Skip to main content
OpenAI

OpenAI Pricing Plans & Tiers

AI models for text, image, and code generation via API

AI & MLusage-basedFrom $0.2/mo

Pricing last verified: March 16, 2026

Data compiled by Arthur Jacquemin, Founder & Lead Analyst
Updated March 16, 2026

Pricing Analysis

OpenAI's tiered model pricing strategy creates a hidden complexity: cheaper per-token rates on GPT-5 mini ($0.25/1M input) mask higher real-world costs when accounting for prompt caching—which pays 90% less on cached tokens. Teams running identical queries repeatedly should model cache hit rates before defaulting to the cheapest model.

The jump from GPT-4.1 mini ($0.80/1M input) to GPT-5 mini ($0.25/1M input) represents a 68% price cut, but output tokens nearly double ($3.20 to $2). This inverts the calculus for long-form generation tasks—models that generate verbose responses become economically worse choices despite cheaper input pricing.

O4-mini ($4/1M input) sits at a critical inflection point: 16x GPT-5 mini's input cost but serves latency-sensitive workloads where smaller models require multi-turn loops. Organizations must track per-request cost-of-delay versus token cost to justify the premium.

Strengths

  • Prompt caching at $0.025 per cached input token on GPT-5 mini enables sub-1-cent economics on frequently-executed queries—a hidden efficiency advantage for batch operations and retrieval-augmented generation.
  • Free tier includes 1GB storage for ChatKit, unusual for API-first services; teams prototyping multi-turn agents avoid cold-start costs.
  • Sora-2 integration at $0.10/sec for video generation creates a unified API for multimodal applications without switching vendors.

Considerations

  • Usage-based pricing offers no guardrails—a misconfigured loop can cost $1,000+ in hours. Enterprise teams need spend governance middleware before OpenAI access.
  • Output token pricing (10x input on GPT-5.4) penalizes summarization and structured extraction tasks; competitors' per-request models may be cheaper for these workloads.
  • No monthly commits or volume discounts; predictability requires custom contracts for >$10K/month spend.
Ideal For

Early-stage AI startups and enterprises with <$5K/month API spend who need maximum model optionality and low per-token costs.

Pricing Takeaway

The cheapest token price ($0.025 cached on GPT-5 mini) is a trap—real savings come from understanding your cache hit ratio and output-token penalty.

Best choice: OpenAI

Try OpenAI free

Pricing Plans (14)

GPT-5.4

$3/mo
Start with GPT-5.4

GPT-5 mini

$0/mo
Start with GPT-5 mini

GPT-4.1

$3/mo
Start with GPT-4.1

GPT-4.1 mini

$1/mo
Start with GPT-4.1 mini

GPT-4.1 nano

$0/mo
Start with GPT-4.1 nano

o4-mini

$4/mo
Start with o4-mini

Realtime API

$2/mo
Start with Realtime API

Sora Video API

Custom
Start with Sora Video API

Image Generation API

$2/mo
Start with Image Generation API

Responses API

Custom
Start with Responses API

Chat Completions API

Custom
Start with Chat Completions API

Assistants API

Custom
Start with Assistants API

Built-in tools

Custom
Start with Built-in tools

AgentKit

Custom
Start with AgentKit

How does OpenAI pricing compare?

See how OpenAI's 14 pricing plans stack up against similar AI & ML tools.

Frequently Asked Questions

How much does OpenAI's API cost per 1,000 tokens?
OpenAI charges per-token usage: GPT-4o costs $0.0025 per 1K input tokens and $0.01 per 1K output tokens. GPT-4o mini costs $0.00015 per 1K input and $0.0006 per 1K output. A typical 5,000-word customer support response using GPT-4o mini costs under $0.01. A 100-message customer support chatbot using mini costs $1-2/month. There's no monthly minimum.
Should I use GPT-4o or GPT-4o mini for production?
GPT-4o mini costs 66x less per token than GPT-4o ($0.00015 vs $0.0025 input). For customer support chatbots, summarization, and classification, mini is 95% as capable and saves 99% of costs. Use GPT-4o for complex reasoning, research synthesis, or code generation where quality justifies the cost.
Does OpenAI charge for API access or just token usage?
OpenAI charges only for token usage — no monthly subscription, API fees, or setup costs. You pay purely based on tokens consumed. This makes OpenAI cheaper than Anthropic's Claude API ($0.08/$0.24 per 1M tokens for Claude 3.5 Sonnet) for equivalent quality tasks.
What is OpenAI's rate limit and does it cost extra?
Free trial accounts have strict rate limits (3 requests/minute). Paid accounts tier by usage: entry level is 3,500 RPM (requests per minute). Higher rate limits are free if you've paid $100+. There's no charge to increase RPM — it's based on historical spending.
Can I cache prompts to reduce API costs on OpenAI?
Yes, OpenAI's prompt caching feature stores up to 128K tokens of input. Cached tokens cost 90% less than regular tokens ($0.00003 vs $0.00015 for mini input). For applications asking the same document repeatedly (customer service FAQs), caching can reduce costs by 50-70%.
How does OpenAI pricing compare to using Claude?
OpenAI GPT-4o mini ($0.00015 input) is 50-60% cheaper than Claude 3.5 Haiku ($0.00040 input). For high-volume, cost-sensitive applications like customer support, OpenAI is significantly cheaper. Claude excels at long-form reasoning where output tokens justify higher per-token costs.

Track OpenAI Pricing Changes

Get notified when pricing changes for this tool and others you follow.

Reviews

No reviews yet. Be the first to review this tool.

Sources

  1. OpenAI Official PricingVendor pricing page
  2. OpenAI ReviewsIndependent reviews on G2
  3. OpenAI ReviewsIndependent reviews on TrustRadius

Are you the team behind OpenAI?

Claim your profile to add custom descriptions, featured badges, and direct demo links.

Claim Your Profile

Related Articles