OpenAI model pricing detail

GPT-5.4 pricing starts at $2.50 input and $15.00 output per 1M tokens.

This page answers the GPT-5.4 pricing query directly. It keeps the current short-context, long-context, and batch rows in one view, then shows the fit and hosted-tool signals that decide whether the flagship row is actually worth paying for.

Current state

This row only matters if the workflow still fits the model.
Live pricing brief
OpenAI currently prices GPT-5.4 by context tier, with a higher long-context row and a separate batch row for short-context processing.

Model in focus

GPT-5.4

Last checked

March 12, 2026

Current source note

This page keeps the literal GPT-5.4 search term because it is the target query. Trust the linked OpenAI pricing and model pages as the source of record if naming or listing changes later.

Price rows

Read the GPT-5.4 rows directly before comparing them to a cheaper model.

The pricing page lists three live GPT-5.4 rows that matter operationally: short context, long context, and short-context batch.

ModeContextInputCached inputOutputDecision read
StandardShort$2.50 per 1M tokens$0.25 per 1M tokens$15 per 1M tokensThis is the lowest live GPT-5.4 row for direct API usage before hosted tools are added.
StandardLong$5.00 per 1M tokens$0.50 per 1M tokens$23 per 1M tokensThe long-context row is materially more expensive and should be priced separately from short-context use.
BatchShort only$1.25 per 1M tokens$0.13 per 1M tokens$7.50 per 1M tokensBatch only applies to the short-context row. There is no separate long-context batch row listed today.

Fit signals

The flagship row earns its keep only when the workload uses the flagship fit.

Context size and tool support are what usually keep GPT-5.4 in the path after the initial price comparison starts.

GPT-5.4 keeps the widest published context window in the current summary rows.

The current OpenAI limits summary lists GPT-5.4 at 1,048,576 tokens. If the workload actually uses that headroom, the row is competing on fit, not just on token price.

The built-in tool set is broader than the cheaper mini path.

OpenAI currently lists GPT-5.4 with Functions, web search, file search, skills, image generation, code interpreter, MCP. If the workflow depends on that breadth, a cheaper model row is not a full replacement.

Long-context use changes the price comparison immediately.

GPT-5.4 short and GPT-5.4 long are not interchangeable estimates. If the workload lands on the long-context row, the flagship path needs its own explicit price read before a cheaper alternative is considered.

Workload snapshots

A GPT-5.4 estimate changes shape quickly once context and hosted tools enter the path.

These snapshots keep the arithmetic simple so the row can be understood before a broader calculator is opened.

Short-context token-only run
10M input tokens and 1M output tokens on the standard short-context row.

Estimate

$40 in base model spend

This is the cleanest GPT-5.4 baseline before file search, web search, or runtime costs enter the path.

Long-context review path
10M input tokens and 1M output tokens on the long-context row.

Estimate

$73 in base model spend

The moment the workload needs the published long-context row, GPT-5.4 stops being comparable to a short-context-only estimate.

Tool-heavy operating path
10M input tokens, 1M output tokens, 40K file-search calls, and a 30 GB average vector-store footprint for 30 days.

Estimate

~$260 before web search or container runtime

The base long-context row is about $73 here, while file search alone adds about $100 in calls plus roughly $87 in storage after the first 1 free GB.

When this row is misleading

The GPT-5.4 token row is not the full operating cost by itself.

These are the three places where the flagship row often gets over-trusted in a budget discussion.

File search can outweigh the model swap.

File search storage bills at $0.10 per GB per day after the first 1 free GB, and file-search calls bill at $2.50 per 1K calls. That often matters more than shaving a few dollars off the token row.

Web search adds its own call and token rules.

If the workflow uses OpenAI web search, the estimate depends on the search path and its separate call pricing, not just on GPT-5.4 input and output tokens.

Container runtime can keep the bill high even after a model swap.

If code interpreter or runtime-heavy workflows stay in the path, the runtime line remains a separate cost surface rather than a hidden part of the GPT-5.4 token row.

Official sources

Check the OpenAI pages behind these pricing rows.

This page keeps the source set narrow so a pricing row can stay auditable instead of drifting into summary-only advice.

Pricing

OpenAI API pricing

Source of record for GPT-5.4 short, long, and batch token pricing.

Open official page
Model page

GPT-5.4 model page

Source of record for the published context window, max output tokens, and built-in tool coverage that change the value of the pricing row.

Open official page

Continue the site

Keep moving through the decision from here.

Use the groups below to move laterally through the decision, not back out into another doc hunt.

Related pages

Stay in the same decision neighborhood instead of backing out to search.

Pricing / Costs

Model pricing, hosted-tool costs, and fit constraints that materially change the operating estimate.

Open page

GPT-5 mini pricing

Single-model pricing brief for GPT-5 mini across standard and batch rows.

Open page

GPT-5.4 context and tool support

Limits brief for GPT-5.4 versus GPT-5 mini context windows, output caps, and tool support.

Open page

Compare pages

Open the pages that turn this topic into a side-by-side decision.

GPT-5.4 vs GPT-5 mini

Side-by-side comparison of GPT-5.4 and GPT-5 mini across price, fit, and tool pressure.

Open page

Cheapest OpenAI model for extraction

Scenario recommendation page for choosing the cheapest workable OpenAI extraction model.

Open page

Replacement pages

Use the likely substitutes, migration targets, or fallback choices as the next click.

GPT-5 mini pricing

Single-model pricing brief for GPT-5 mini across standard and batch rows.

Open page

OpenAI API pricing calculator

Interactive calculator for model tokens, hosted tools, and runtime in one estimate.

Open page

OpenAI file search pricing

Tool-cost brief for file search pricing across storage, tool calls, and model-token exposure.

Open page

Source category pages

Trace the source families behind this page instead of opening random docs in isolation.

Pricing sources

Official pricing pages used to support model, tool-cost, and calculator estimates.

Open page

Model sources

Official model pages used for context windows, output caps, and built-in tool coverage.

Open page

Return

Return to the OpenAI tracker

Go back to the main OpenAI decision surface to compare this pricing brief against hosted-tool costs, deprecation risk, and the broader family matrix.

Back to OpenAI tracker