OpenAI model pricing detail

GPT-5 mini pricing starts at $0.25 input and $2.00 output per 1M tokens.

This page answers the GPT-5 mini pricing query directly. It keeps the live standard and batch rows in one view, then shows where the cheap row is a real answer and where context or tool fit still pushes the decision upward.

Current state

This row only matters if the workflow still fits the model.
Live pricing brief
OpenAI currently lists GPT-5 mini as a low-cost general-purpose row with both standard and batch pricing.

Model in focus

GPT-5 mini

Last checked

March 12, 2026

Price rows

Read the cheap row directly before assuming it fits every workload.

GPT-5 mini looks simple on the pricing page, but the real question is whether the low row survives context, tool, and hosted-tool pressure.

ModeContextInputCached inputOutputDecision read
StandardStandard$0.25 per 1M tokens$0.03 per 1M tokens$2.00 per 1M tokensThis is the baseline live GPT-5 mini row for direct API usage before hosted tools are added.
BatchStandard$0.13 per 1M tokens$0.01 per 1M tokens$1.00 per 1M tokensBatch is the cheapest published GPT-5 mini path for large-volume extraction or classification work.

Fit signals

The cheap row works best when the workflow really is small, repeatable, and tool-light.

GPT-5 mini becomes the right answer when the workload stays inside the published context ceiling and does not depend on the broader flagship tool set.

GPT-5 mini is cheap because it is narrower, not because it is a universal substitute.

The current OpenAI limits summary lists GPT-5 mini at 400,000 tokens. That is still large, but it is materially smaller than the flagship path and should be treated as a real fit constraint.

The published tool set is narrower than GPT-5.4.

OpenAI currently lists GPT-5 mini with Functions, web search, file search, MCP. If the workflow needs a broader tool surface, the low token row can still be the wrong operating choice.

Batch pricing is where GPT-5 mini becomes especially hard to ignore.

When the workload is high-volume and repeatable, the batch row can matter more than the standard row. That makes GPT-5 mini a serious extraction or classification baseline even before more aggressive cheap rows are considered.

Workload snapshots

GPT-5 mini becomes compelling when the workload is cheap and well-bounded on purpose.

These snapshots keep the arithmetic simple so the mini row can be judged without opening the full model matrix first.

Short-turn token-only work
10M input tokens and 1M output tokens on the standard row.

Estimate

$4.50 in base model spend

This is the cleanest mini baseline for extraction, classification, or short-turn generation without hosted tools.

High-volume batch extraction
100M input tokens and 10M output tokens on the batch row.

Estimate

$23 in base model spend

If the workload is repeatable enough for batch, GPT-5 mini becomes a strong default cost anchor.

Long-document extraction boundary
A single extraction path that grows toward the published 400,000-token context ceiling.

Estimate

Cheap on paper, but only if the document still fits the published ceiling

Once the prompt or retrieval context outgrows the mini ceiling, the low row stops being a live option instead of merely becoming less efficient.

When this row is misleading

The cheapest visible token row is not always the cheapest working path.

These are the places where GPT-5 mini usually gets over-credited in a budget conversation.

Hosted tools can flatten the model savings.

File search, web search, and container runtime can stay expensive even when the base model row gets much cheaper. A model swap is not the same thing as a full workflow optimization.

The smaller context window can turn the cheap row into a dead end.

If the extraction or reasoning path depends on longer prompts, larger retrieval payloads, or richer context carry-forward, GPT-5 mini should not be priced as if it were a drop-in replacement.

Tool-light assumptions can hide a real fit mismatch.

If the workflow needs built-in image generation, code interpreter, or skills on the same model path, the cheap GPT-5 mini row no longer answers the whole decision by itself.

Official sources

Check the OpenAI pages behind these pricing rows.

This page keeps the source set narrow so a pricing row can stay auditable instead of drifting into summary-only advice.

Pricing

OpenAI API pricing

Source of record for GPT-5 mini standard and batch token pricing.

Open official page
Model page

GPT-5 mini model page

Source of record for GPT-5 mini context window, max output tokens, and current built-in tool support.

Open official page

Continue the site

Keep moving through the decision from here.

Use the groups below to move laterally through the decision, not back out into another doc hunt.

Related pages

Stay in the same decision neighborhood instead of backing out to search.

Pricing / Costs

Model pricing, hosted-tool costs, and fit constraints that materially change the operating estimate.

Open page

GPT-5.4 pricing

Single-model pricing brief for GPT-5.4 across short, long, and batch rows.

Open page

GPT-5.4 context and tool support

Limits brief for GPT-5.4 versus GPT-5 mini context windows, output caps, and tool support.

Open page

Compare pages

Open the pages that turn this topic into a side-by-side decision.

GPT-5.4 vs GPT-5 mini

Side-by-side comparison of GPT-5.4 and GPT-5 mini across price, fit, and tool pressure.

Open page

Cheapest OpenAI model for extraction

Scenario recommendation page for choosing the cheapest workable OpenAI extraction model.

Open page

Replacement pages

Use the likely substitutes, migration targets, or fallback choices as the next click.

GPT-5.4 pricing

Single-model pricing brief for GPT-5.4 across short, long, and batch rows.

Open page

OpenAI API pricing calculator

Interactive calculator for model tokens, hosted tools, and runtime in one estimate.

Open page

OpenAI file search pricing

Tool-cost brief for file search pricing across storage, tool calls, and model-token exposure.

Open page

Source category pages

Trace the source families behind this page instead of opening random docs in isolation.

Pricing sources

Official pricing pages used to support model, tool-cost, and calculator estimates.

Open page

Model sources

Official model pages used for context windows, output caps, and built-in tool coverage.

Open page

Return

Return to the OpenAI tracker

Go back to the main OpenAI decision surface to compare GPT-5 mini against current tool costs, lifecycle pressure, and the wider family matrix.

Back to OpenAI tracker