Estimate LLM API costs for your team
Estimate and compare pricing for world-class language models. Our precision-engineered engine factors in context tokens, multimodal inputs, and volume discounts to provide a transparent financial roadmap for your AI infrastructure.
Select a model to see the estimated cost
Build a unified AI workspace for your team
- Access all AI models in one place
- Manage usage, costs & permissions per team
- Connect your own API keys
- ...and much more
Try free
Trusted by 20,641+ customers
Context400k
Input/1M$0.20
Output/1M$1.25
Context400k
Input/1M$0.75
Output/1M$4.50
Context1.1M
Input/1M$2.50
Output/1M$15.00
Context1.1M
Input/1M$30.00
Output/1M$180.00
Context128k
Input/1M$1.75
Output/1M$14.00
Context400k
Input/1M$1.75
Output/1M$14.00
Context128k
Input/1M$1.75
Output/1M$14.00
Context1.0M
Input/1M$3.00
Output/1M$15.00
Context1.0M
Input/1M$5.00
Output/1M$25.00
Context400k
Input/1M$1.75
Output/1M$14.00
Context400k
Input/1M$21.00
Output/1M$168.00
Context128k
Input/1M$1.75
Output/1M$14.00
Context400k
Input/1M$1.75
Output/1M$14.00
Context1.0M
Input/1M$0.25
Output/1M$1.50
Context131k
Input/1M$0.25
Output/1M$60.00
Context1.0M
Input/1M$2.00
Output/1M$12.00
Context1.0M
Input/1M$2.00
Output/1M$12.00
Context400k
Input/1M$1.25
Output/1M$10.00
Context400k
Input/1M$0.25
Output/1M$2.00
Context400k
Input/1M$1.25
Output/1M$10.00
Context400k
Input/1M$1.25
Output/1M$10.00
Context128k
Input/1M$1.25
Output/1M$10.00
Context1.0M
Input/1M$0.00
Output/1M$0.00
Context1.0M
Input/1M$0.00
Output/1M$0.00
Context262k
Input/1M$0.40
Output/1M$2.00
Context400k
Input/1M$0.00
Output/1M$0.00
Context1.0M
Input/1M$1.00
Output/1M$3.00
Context400k
Input/1M$0.00
Output/1M$0.00
Context205k
Input/1M$0.30
Output/1M$1.20
Context262k
Input/1M$0.15
Output/1M$0.60
Context200k
Input/1M$1.20
Output/1M$4.00
Context2.0M
Input/1M$2.00
Output/1M$6.00
Context2.0M
Input/1M$2.00
Output/1M$6.00
Context262k
Input/1M$0.10
Output/1M$0.50
Context262k
Input/1M$0.00
Output/1M$0.00
Context262k
Input/1M$0.00
Output/1M$0.00
Context1.1M
Input/1M$2.50
Output/1M$15.00
Context1.1M
Input/1M$30.00
Output/1M$180.00
Context400k
Input/1M$15.00
Output/1M$120.00
Context200k
Input/1M$5.00
Output/1M$25.00
Context1.0M
Input/1M$0.25
Output/1M$1.50
Context128k
Input/1M$0.25
Output/1M$0.75
Context400k
Input/1M$1.75
Output/1M$14.00
Context1.0M
Input/1M$2.00
Output/1M$12.00
Context1.0M
Input/1M$2.00
Output/1M$12.00
Context1.0M
Input/1M$3.00
Output/1M$15.00
Context262k
Input/1M$0.60
Output/1M$3.60
Context1.0M
Input/1M$0.40
Output/1M$2.40
Context400k
Input/1M$1.25
Output/1M$10.00
Context203k
Input/1M$1.00
Output/1M$3.20
Context205k
Input/1M$0.30
Output/1M$1.20
…
Frequently Asked Questions
What is a token?
A token is a chunk of text — roughly ¾ of a word in English. For example, "chatbot" is two tokens. Pricing is based on the number of tokens sent (input) and received (output). For programmatic tokenization, see the Tiktoken library.
How is the cost calculated?
We estimate daily token usage based on your team size, chats per user, response length, and image/PDF uploads. Tokens are split 40% input / 60% output to reflect typical conversations. The token count is then multiplied by each model's per-million-token price. When multiple models are selected, usage is distributed evenly across them.
Why are input and output priced differently?
LLM providers charge separately for input tokens (your prompts, uploaded documents, images) and output tokens (the model's responses). Output tokens are typically more expensive because they require more compute to generate.
What does "Add to Total Cost" do?
It adds the model to your cost estimate. When you select multiple models, the tool assumes your team's usage is split evenly across them — so each model handles a proportional share of chats, images, and documents. The total is the sum of all models' shares.
Can I compare costs across different providers?
Yes. Use the provider tabs to filter by vendor, or search for specific models. Add multiple models to see a side-by-side cost breakdown in the left panel. You can also sort by price to quickly find the cheapest or most expensive options.
What does the "Capabilities" sort do?
It ranks models by a composite score that factors in feature support (attachments, reasoning, tool calling, temperature control), the number of input/output modalities, how recently the model was released or updated, and whether it comes from a major provider.
What is a context window?
The context window is the maximum number of tokens a model can process in a single request — including both your input and the model's output. Larger context windows let you send longer documents or maintain longer conversation histories, but may cost more.
How accurate is this estimation?
This tool provides a ballpark estimate. Actual costs may differ due to prompt caching, batched API calls, volume discounts, reasoning token overhead, and provider-specific billing rules. Use it for budgeting and comparison, not as an invoice prediction.










