OpenAI tool-cost detail

OpenAI container pricing starts when code interpreter runtime starts.

This page answers the container-pricing question directly. It keeps the current container rate, the runtime trigger, and realistic workload patterns in one source-linked view so code interpreter cost does not disappear behind the model row.

Current state

This tool has more than one meter, and each one can dominate a different workload.
Live cost brief
OpenAI currently prices code interpreter usage at $0.03 per container per 20 minutes, and that runtime bill sits on top of normal model-token charges.

Last checked

March 12, 2026

Container rate

$0.03 per container per 20 minutes

This is the current published runtime line item for code interpreter containers.

Billing trigger

Runtime starts billing when a container is needed for code execution

The container line is separate from the base model row and should be estimated separately from tokens.

Stacking behavior

Container cost adds to model spend and any other hosted-tool charges

A cheap model row does not absorb runtime cost. Container pricing is its own operating meter.

Cost anatomy

Container pricing is a runtime meter, not a token discount or token surcharge.

If code execution happens in the path, the container line should be estimated alongside tokens rather than treated as a hidden part of the model row.

The container line is its own billable runtime meter.
OpenAI currently lists code interpreter usage at $0.03 per container per 20 minutes. That means code execution creates a separate runtime line item before model tokens are even compared.
Containers do not replace model-token billing.
The container meter is additive. Teams still pay for the chosen model path, and any retrieval or search tools in the same workflow still keep their own charges.
Runtime-heavy automation gets expensive by repetition, not by one session.
Single interactive runs look cheap. Repeated containers across workflows, analyst jobs, or automation turns are what turn this into a real budget line.

Workload examples

The runtime bill stays quiet until container count starts repeating.

These examples isolate the runtime line so the container meter can be seen before model tokens or retrieval costs are layered back in.

ScenarioWorkloadRuntime meterModel overlapThird meterDecision readSources
Low-frequency analyst sessions20 code-interpreter sessions during the month, one container each.20 x $0.03 per container per 20 minutes = $0.60Model tokens still matter, but runtime is small enough that the container line stays visible and manageable.No extra hosted-tool assumption in this example. This is the cleanest runtime-only read.At low frequency, container pricing is a visible but small add-on meter.
Sustained multi-session operations400 code-interpreter turns during the month, one container each.400 x $0.03 per container per 20 minutes = $12.00Runtime still looks modest, but it now repeats often enough that the container line should be priced explicitly beside the model row.If file search or web search also sits in the path, those meters stack separately.Once the workflow runs all day, runtime stops being background noise and becomes a repeatable operations cost.
Container-heavy automation25,000 code-execution jobs during the month, one container per job.25,000 x $0.03 per container per 20 minutes = $750.00Model tokens now sit underneath a large runtime line, so changing the model alone no longer solves the budget question.Additional file search or web search usage would stack on top of the runtime bill rather than replacing it.At automation scale, container count becomes the budget question before headline model pricing does.

Control levers

The best container optimization is usually to reduce needless runtime starts.

Code interpreter cost control comes from how often runtime is invoked and whether it belongs in the path at all.

Route only the turns that truly need runtime.

If a turn only needs structured text generation, keep it off the code interpreter path. Containers should be reserved for genuine execution, analysis, or file-processing steps.

Price runtime and tokens separately in every estimate.

A model swap can still help, but only after the container line has been isolated. Otherwise teams over-credit the cheaper model row for savings it cannot produce.

Check repeat usage before chasing micro-optimizations.

The real cost jump comes from repeated container starts across sessions or jobs. Start by measuring how often runtime is invoked before tuning prompts or token payloads.

Decision signals

What usually determines whether container pricing matters yet.

Use these signals to decide whether code interpreter is just an add-on or already the main cost risk in the workflow.

Interactive use keeps the container line small.

If code execution is rare, container pricing is a visible but modest add-on. The model row and other tools still drive most of the estimate.

Operational repetition makes runtime the first question.

Frequent analysis turns or automation jobs make container count the real budget driver long before model-token differences fully show up.

Runtime is often the hidden reason model savings disappoint.

If the workflow still starts the same number of containers, a cheaper model row may only trim a small part of the total bill.

Official sources

Check the OpenAI pages behind these cost lines.

This page keeps the source set narrow so the cost brief can stay auditable instead of drifting into guesswork.

Pricing

OpenAI API pricing

Source of record for the current code interpreter container price and the effective billing unit.

Open official page
Guide

Code interpreter guide

Documents the hosted tool path and runtime behavior that turn code interpreter usage into a container-cost question instead of a pure model-token question.

Open official page

Continue the site

Keep moving through the decision from here.

Use the groups below to move laterally through the decision, not back out into another doc hunt.

Related pages

Stay in the same decision neighborhood instead of backing out to search.

Pricing / Costs

Model pricing, hosted-tool costs, and fit constraints that materially change the operating estimate.

Open page

OpenAI file search pricing

Tool-cost brief for file search pricing across storage, tool calls, and model-token exposure.

Open page

OpenAI web search pricing

Tool-cost brief for web search pricing across standard and preview search paths.

Open page

Compare pages

Open the pages that turn this topic into a side-by-side decision.

GPT-5.4 vs GPT-5 mini

Side-by-side comparison of GPT-5.4 and GPT-5 mini across price, fit, and tool pressure.

Open page

Cheapest OpenAI model for extraction

Scenario recommendation page for choosing the cheapest workable OpenAI extraction model.

Open page

Replacement pages

Use the likely substitutes, migration targets, or fallback choices as the next click.

OpenAI API pricing calculator

Interactive calculator for model tokens, hosted tools, and runtime in one estimate.

Open page

GPT-5 mini pricing

Single-model pricing brief for GPT-5 mini across standard and batch rows.

Open page

GPT-5.4 pricing

Single-model pricing brief for GPT-5.4 across short, long, and batch rows.

Open page

Source category pages

Trace the source families behind this page instead of opening random docs in isolation.

Pricing sources

Official pricing pages used to support model, tool-cost, and calculator estimates.

Open page

Guide and API reference sources

Operational guides and API references used by tool-cost, migration, and calculator pages.

Open page

Return

Return to the OpenAI tracker
Go back to the main OpenAI decision surface to compare this runtime cost with current model pricing, other hosted-tool charges, and lifecycle risk.