Docs Home | Agent-Native | Benchmark Report | Framework Comparison | Ecosystem | API | Configuration | Examples
This page explains what benchmark scripts measure, what they do not measure, and how to interpret results safely.
bun run bench:agentMeasures end-to-end path cost for:
- direct parse (
processText) AgentBridgetool path- MCP
tools/callhandler path
bun run bench:frameworksMeasures local tool-dispatch overhead with one shared handler implementation:
- Direct handler baseline
- Qirrel AgentBridge
- Qirrel MCP handler
- LangChain
tool()(if installed) - AI SDK
tool()(if installed)
bun run bench:reportGenerates/updates:
- Benchmarks are local machine measurements, not universal truth.
- This suite isolates orchestration overhead (no external model API calls in framework comparison track).
- Optional frameworks are skipped if dependencies are unavailable.
ops/sec: higher is better.avg ms: lower is better.p99 ms: lower is better for tail latency stability.vs direct: slowdown relative to direct baseline (1.00x= equal to baseline).
- Close resource-heavy apps.
- Run each benchmark at least twice.
- Compare runs on the same runtime version and machine profile.
- Commit results only when stable.
- A faster wrapper in synthetic benchmarks does not automatically mean better full-system latency.
- Cross-machine comparisons are usually invalid unless hardware/runtime are normalized.
- Throughput alone is not enough; include ergonomics, interoperability, and failure handling in selection decisions.