Skip to content

Commit 75ca893

Browse files
committed
docs: add Costs section to README with token usage and cost breakdown
Adds a new ## Costs section summarizing observed LLM costs based on maintainer testing (~82 requests, 246k tokens). Includes a stats table (avg, 95th percentile, peak), a token breakdown table, and a key insight explaining that output tokens drive ~50% of cost despite being only 17% of volume. Includes a disclaimer that figures are based on personal testing. https://claude.ai/code/session_01Raf7fKQax1GZEFoYUsGyxD
1 parent 18c735e commit 75ca893

1 file changed

Lines changed: 27 additions & 0 deletions

File tree

README.md

Lines changed: 27 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -488,6 +488,33 @@ The PR comment is always posted — with a GIF if recording succeeded, or screen
488488

489489
---
490490

491+
## Costs
492+
493+
> **Disclaimer:** The figures below are based on the maintainer's own testing and may not reflect your usage patterns. Token consumption varies significantly by diff size, app complexity, and retry count. Use these numbers as a rough reference only — your costs may be higher or lower.
494+
495+
Each GitGlimpse run makes one or more LLM calls to generate the Playwright interaction script. The table below summarizes observed costs using `claude-sonnet-4-6` with default settings.
496+
497+
| Metric | Value |
498+
|---|---|
499+
| Average cost per request | ~$0.015 (1.5¢) |
500+
| 95th percentile | ~$0.035 |
501+
| Peak (max observed) | ~$0.047 |
502+
| Requests per USD | ~66 |
503+
| Sample size | ~82 requests / 246k tokens |
504+
505+
**Token breakdown (per request, approximate):**
506+
507+
| Token type | Volume | Share of tokens | Share of cost |
508+
|---|---|---|---|
509+
| Input | ~205k total | ~83% | ~50% |
510+
| Output | ~42k total | ~17% | ~50% |
511+
512+
**Key insight:** Output tokens drive roughly **50% of the cost** despite representing only **17% of the token volume** — a direct consequence of output tokens being priced ~5× higher than input tokens in most Claude pricing tiers. Reducing script verbosity or capping `maxDuration` can lower output token counts and trim costs noticeably.
513+
514+
For high-frequency teams, use `trigger.mode: 'smart'` or `'on-demand'` to avoid running the LLM on every push.
515+
516+
---
517+
491518
## Development
492519
493520
```bash

0 commit comments

Comments
 (0)