LLM Token Cost Calculator
Compare Claude, ChatGPT, and Gemini API costs at any volume. Same prompt shape, three price tags.
Estimate API spend per request
INTERACTIVE · LIVE · VERIFIED PRICESEstimate one request. Cache rate captures stable prefixes (system prompts, long docs); only Claude has explicit prompt caching shown here. Cards sort cheapest → most expensive automatically.
Cache math: when the cache slider is non-zero, that fraction of input tokens is billed at 10% of the model's input rate (the standard prompt-caching read multiplier). The other models in the list don't get a cache discount here — partly because their caching APIs are priced and exposed differently, partly to keep this view honest. For the deeper Claude prompt-caching breakdown (write multipliers, TTLs, Batch API stacking), see the full Claude review.