# Loki Mode Benchmark Results
**Generated:** 2026-01-05 09:31:14
## Overview
This directory contains benchmark results for Loki Mode multi-agent system.
## Methodology
Loki Mode uses its multi-agent architecture to solve each problem:
1. **Architect Agent** analyzes the problem
2. **Engineer Agent** implements the solution
3. **QA Agent** validates with test cases
4. **Review Agent** checks code quality
This mirrors real-world software development more accurately than single-agent approaches.
## Running Benchmarks
```bash
# Setup only (download datasets)
./benchmarks/run-benchmarks.sh all
# Execute with Claude
./benchmarks/run-benchmarks.sh humaneval --execute
./benchmarks/run-benchmarks.sh humaneval --execute --limit 10 # First 10 only
./benchmarks/run-benchmarks.sh swebench --execute --limit 5 # First 5 only
# Use different model
./benchmarks/run-benchmarks.sh humaneval --execute --model opus
```
Source: claude-code-templates (MIT). See About Us for full credits.