Skip to main content
tokenmath
Menu

GPT-5 Nano token & cost calculator

OpenAI GPT-5 Nano is the cheapest entry in the GPT-5 family — a model built specifically for the workloads where you previously couldn't justify calling an LLM at all. At $0.05 input / $0.40 output per million tokens, Nano sits at roughly 1/25th the cost of full GPT-5 and well below Claude 4.5 Haiku or Gemini 2.5 Flash. The implication for product design is real: tasks that were uneconomic at GPT-4-class pricing become straightforwardly affordable here.

The honest framing for Nano: it shifts the bottleneck from cost to reliability. When the model bill is fractions of a cent per request, the engineering question becomes "how often does this model get the right answer," not "can we afford to call it." Spend your time on evals, schema validation, and routing logic; the per-million numbers will take care of themselves.

Client-side. Never uploaded.
0 / 1,000,000 charactersContext window: 400,000 tokens
Or start with an example
Total estimated cost
<$0.01GPT-5 Nano
Tokensexact
0
Input cost
$0.00
Output cost (est.)
<$0.01
@ 1,024 response tokens
Context used
0%
of 400,000
Verified 2026-05-09 · exact
Saved scenariosnone yet

Saved on this browser only — never uploaded. Up to 10 scenarios.

Tip: save a scenario when you have a prompt + model + response length you might revisit. Useful for sizing features before committing to a vendor.

Verify privacysince this page loaded — updates live
Prompt uploads0Always 0 — by design
Outgoing requests0Analytics + page assets only — no prompt content
Cookies on this origin0Vercel Analytics + Clarity may set first-party cookies
localStorage keys0Theme preference + saved scenarios live here
Server endpoints1/api/og only — accepts title + subtitle, never prompt text
Inspect

Open DevTools → Network. Type into the calculator. No request bodies should contain your prompt text.

Pricing

Nano is flat-priced. Even the most input-heavy realistic workload at this rate still fits comfortably under a cent per request.

TierInput $/MOutput $/M
All input$0.05$0.4
Context window400,000 tokens

Verified against openai.com on 2026-05-09.

Worked examples

The scenarios below show what Nano costs at three typical prompt sizes. All three round to a fraction of a cent — the price floor is the point.

ScenarioInputOutputCost
Short chat turn
A typical Q&A turn with a small system prompt.
800400<$0.01
System prompt + tool spec
A larger context window with a tool schema, single response.
5,000500<$0.01
Long document Q&A
A long-form input (e.g. transcript) with a structured response.
50,0001,500<$0.01

A pattern worth knowing: Nano + a downstream classifier is often better than Mini. For high-volume classification or extraction, calling Nano with a strict JSON schema and a tight max_tokens cap, then validating the response with deterministic code, is frequently more reliable than running Mini at default settings. The cheap model with rigid output constraints often beats a smarter model with sloppy ones.

How is this counted?

Nano uses OpenAI's canonical o200k_base tokenizer, the same vocab as GPT-5 and GPT-4.1. The count is exact via gpt-tokenizer (MIT) — calibration factor 1.0. Inputs over 50,000 characters run in a Web Worker.

FAQ

What is Nano good for?
Volume tasks where you'd historically have used a small fine-tuned model: bulk classification, sentiment scoring, intent detection, lightweight extraction, schema validation, A/B-test variant generation, batch rewrites. Nano is fast and cheap enough that the model bill becomes a rounding error in your infrastructure spend.
When does Nano stop being good enough?
Anything that requires reliable multi-step reasoning, long-document synthesis, or generating customer-facing prose at high quality. Nano is not the right model for code generation in production, complex summarization, or any task where a confident-but-wrong answer is worse than no answer. Promote those workloads to Mini or full GPT-5.
How accurate is the token count?
Exact. The calculator uses OpenAI's canonical o200k_base tokenizer (via gpt-tokenizer). No approximation, no calibration.
Is Nano really 25× cheaper than GPT-5?
On input, yes — $0.05/M for Nano vs $1.25/M for GPT-5. On output, the gap is similar: $0.40/M vs $10/M. At Nano prices a million-token request costs less than a single GPT-5 short chat turn. The tradeoff is in capability, not in the math.
Can I batch Nano calls?
Yes, and you should. OpenAI's Batch API discounts most models 50% for jobs that can wait up to 24 hours; on Nano, that pushes cost into territory where the model bill genuinely disappears from your P&L. The calculator above shows synchronous-API pricing — divide by two for batch.

Compare against every other model

To see this exact prompt scored against every supported model, sorted by total cost, paste it into the home calculator and toggle Compare across all models. At Nano's price floor the comparison is mostly about capability fit, not dollars.

Related models

The most relevant comparisons: GPT-5 Mini (when Nano's capability bar is too low) and the cross-vendor budget tiers — Claude 4.5 Haiku and Gemini 2.5 Flash.

Keyboard shortcuts

Press ? any time to reopen this list.

Show this overlay?
Toggle themet
Focus the prompt textarea/
Go to homegh
Go to modelsgm
Go to pricing datagp
Go to changeloggc
Go to aboutga
Close overlays / dialogsEsc