Startup copilot
Suppose you run 10M input tokens and 2.5M output tokens per month on Claude Sonnet 4. The math is
10 x $3 + 2.5 x $15, which lands at about $67.50 monthly before any batch discounts.
Use this page to answer three practical questions fast: what Anthropic charges per model today, how much your monthly token volume will cost, and when Claude API is a better fit than Claude Max or a comparable OpenAI API workflow.
Claude API pricing is metered on input and output tokens. Pick a model, enter your monthly token estimates, and this calculator will show the directional bill. It also lets you apply Anthropic's published 50% batch-processing discount when that mode is relevant.
Anthropic's pricing page currently lists three public Claude API model tiers that matter for most buyers. Claude Haiku 3.5 is the cheapest option and is usually evaluated for lightweight classification, routing, or internal summarization. Claude Sonnet 4 is the common general-purpose purchase because it stays within a mid-range budget while still being useful for coding, support automation, and product features. Claude Opus 4.1 sits in the premium tier where quality and complex reasoning justify a much larger output-token bill.
The pricing math is simple but easy to underestimate in practice. Teams focus on input cost because prompts are visible, but output tokens often become the larger part of the bill when the model writes long answers, code, or structured documents. That is why you should model both sides of the meter and then compare the result against your expected volume, latency target, and quality threshold before you commit to one model as the default.
| Model | Input price | Output price | Best fit |
|---|---|---|---|
| Claude Haiku 3.5Lower-cost model for lightweight production tasks. | $0.80 / 1M tokens | $4 / 1M tokens | Classification, short summaries, cost-sensitive routing. |
| Claude Sonnet 4Default production tier for many teams. | $3 / 1M tokens | $15 / 1M tokens | Developer copilots, support workflows, AI product features. |
| Claude Opus 4.1Premium reasoning tier. | $15 / 1M tokens | $75 / 1M tokens | High-stakes reasoning, premium coding, slower but higher-value tasks. |
Anthropic's pricing page also documents prompt caching modes. That matters if your workflow reuses long system prompts or big repeated contexts.
Batch jobs currently receive a 50% discount according to Anthropic's pricing page. That can materially change overnight workflows.
Most teams under-model output tokens. If the model writes code, JSON, or long answers, the output side becomes the real budget lever.
This is where a lot of buyers get tripped up. Claude API and Claude Max are not two price tiers of the same product. Claude API is token-metered and designed for developers, product teams, and automation systems. Claude Max is a paid Claude app plan for humans using the Claude interface more heavily. It is closer to "I work inside Claude every day" than to "I need API throughput for an application."
Anthropic's pricing page currently shows Claude Pro at $20 per month and Claude Max tiers starting at $100 per month, with Max framed around higher usage compared with Pro. That does not mean Max replaces API spend. It means Max may be the better buy for a power user doing manual research, writing, and coding inside the app. If you need to power a product, agent workflow, or internal automation system, Claude API is still the correct commercial model because it gives you the actual programmatic access Max does not include.
The clean way to think about "which is cheaper" is to separate manual usage from integrated usage. If a founder or operator spends most of the day in the Claude app and only occasionally prototypes through code, Max might be a worthwhile human productivity purchase. If the job is embedding Claude into a product, backend service, or agent pipeline, you are comparing model cost per token, not subscription comfort. That is why the most accurate answer is usually: Claude Max for heavy human app usage, Claude API for software, and sometimes both for teams that need both workflows.
Buyers rarely evaluate Claude API pricing in a vacuum. They usually compare it against OpenAI because both vendors serve overlapping coding, chat, automation, and product-integration use cases. The hard part is that price alone does not settle the decision. A cheaper input rate can still lose if output costs spike or if the model fit is weaker for your workload.
| Provider / model | Input | Output | Interpretation |
|---|---|---|---|
| Claude Sonnet 4Anthropic general-purpose tier | $3 / 1M | $15 / 1M | Close to premium production pricing, especially if you like Anthropic workflow quality. |
| Claude Opus 4.1Anthropic premium tier | $15 / 1M | $75 / 1M | Clearly premium. Budget it for high-value outputs, not broad default traffic. |
| Claude Haiku 3.5Anthropic cost-sensitive tier | $0.80 / 1M | $4 / 1M | Useful when budget discipline matters more than top-end capability. |
| OpenAI GPT-5.4OpenAI current standard benchmark | $2.50 / 1M | $15 / 1M | Very close to Sonnet on output cost, slightly lower on input cost. |
| OpenAI GPT-4oOpenAI lower-priced general-purpose reference | $2.50 / 1M | $10 / 1M | Lower output cost than Sonnet, but the decision still depends on capability and workflow fit. |
The most practical buying move is to estimate three workload mixes: low-output classification, balanced assistant calls, and high-output generation. Run all three through the calculator above. That gives you a better directional answer than obsessing over a single headline price because it mirrors how your product will actually use the models in production.
Claude API pricing is Anthropic's token-based billing model for programmatic access to Claude models. Instead of paying a flat monthly subscription for a human-facing app experience, you pay for the tokens your software sends in and the tokens the model returns. That makes Claude API pricing highly flexible, but it also means your bill depends on product behavior, not just plan choice.
For teams shipping software, this is usually the right structure because the billing model maps to real usage. If you deploy a support agent that answers twice as many customer tickets next month, the bill rises with that activity. If you batch jobs overnight or optimize prompts to reduce output length, the bill falls. Claude API pricing is therefore not only a rate card. It is an operating lever that teams can improve through prompt design, routing logic, batching, and caching.
Anthropic's rate-limit documentation is important because the buying decision is not just about raw token price. The docs state that new users receive a small amount of free credits to explore the API, then unlock higher usage tiers through deposits and payment history. In other words, there is exploratory access, but not a permanent unlimited free tier you can build a product around.
Rate limits also shape the real economics of the API. Even if the token price looks acceptable, you still need enough throughput for your requests per minute and enough token capacity for your workload burst profile. That is why a serious buyer should read the pricing page and the rate-limit page together, not separately. Price tells you what usage costs. Rate limits tell you whether the same product can handle your workflow shape.
Suppose you run 10M input tokens and 2.5M output tokens per month on Claude Sonnet 4. The math is
10 x $3 + 2.5 x $15, which lands at about $67.50 monthly before any batch discounts.
If you keep input at 10M tokens but let output grow to 8M tokens, Sonnet rises to roughly
$150. That is why long answers and verbose tool outputs are the budget leak most teams miss early.
A 20M input / 5M output Sonnet workflow would normally cost about $135, but the published 50% batch
discount cuts the directional bill to about $67.50 if your workload qualifies for batch processing.
Anthropic currently lists Claude Opus 4.1 at $15 input and $75 output per million tokens, Claude Sonnet 4 at $3 and $15, and Claude Haiku 3.5 at $0.80 and $4.
No. Claude Max is a paid Claude app plan. It is not API credit, so programmatic usage is still billed separately through Claude API pricing.
Anthropic documents small exploratory API credits for new users, but sustained capacity depends on usage-tier progression and payment history rather than a permanent free plan.
Model the same monthly input and output token volumes across both vendors. Then compare not only price but also quality fit, latency, tooling, and the amount of output your workflow generates.
Because output tokens are often priced higher than input tokens. The more code, JSON, or long-form text the model emits, the faster the bill climbs.
Claude Max makes sense for heavy human usage inside the Claude app. If your main workflow is product integration or agent automation, Claude API remains the right commercial path.