v2/benchmark/README.md
Production-ready benchmarking for Claude Flow with real command execution and authentic metrics.
cd benchmark
pip install -e .
# Run real benchmark
python examples/real_swarm_benchmark.py "Build REST API"
./claude-flow commands via subprocess--non-interactive --output-format stream-json./claude-flow swarm "task" --non-interactive --output-format stream-json
./claude-flow hive-mind spawn "task" --non-interactive
./claude-flow sparc run code "task" --non-interactive
from swarm_benchmark import BenchmarkEngine
engine = BenchmarkEngine(use_real_executor=True)
result = await engine.run_real_benchmark("Build microservices")
print(f"Tokens: {result.metrics['total_tokens']}")
pytest tests/
python examples/verify_real_integration.py
Version: 2.0.0 | Status: Production Ready | Real: Yes | Mocks: None