AI Model Comparison Table

Sort by:

GPT-5.4

OpenAI
2026-03
Context
1.0M
Max Output
128K
Input $/1M
$2.50
Output $/1M
$15.00
MMLU
HumanEval

GPT-5.4 Pro

OpenAI
2026-03
Context
1.0M
Max Output
128K
Input $/1M
$30.00
Output $/1M
$180.00
MMLU
HumanEval

Gemini 3.1 Flash-Lite

Google
2026-03
Context
1.0M
Max Output
66K
Input $/1M
$0.25
Output $/1M
$1.50
MMLU
HumanEval

GPT-5.3-Codex

OpenAI
2026-02
Context
400K
Max Output
128K
Input $/1M
$1.75
Output $/1M
$14.00
MMLU
HumanEval

Claude Opus 4.6

Anthropic
2026-02
Context
200K
Max Output
128K
Input $/1M
$5.00
Output $/1M
$25.00
MMLU
HumanEval

Gemini 3.1 Pro

Google
2026-02
Context
1.0M
Max Output
66K
Input $/1M
$2.00
Output $/1M
$12.00
MMLU
HumanEval

Claude Sonnet 4.6

Anthropic
2026-01
Context
200K
Max Output
64K
Input $/1M
$3.00
Output $/1M
$15.00
MMLU
HumanEval

Gemini 3 Flash

Google
2026-01
Context
1.0M
Max Output
66K
Input $/1M
$0.50
Output $/1M
$3.00
MMLU
HumanEval

Grok 4.1 Fast

xAI
2026-01
Context
2.0M
Max Output
33K
Input $/1M
$0.20
Output $/1M
$0.50
MMLU
HumanEval

DeepSeek Chat (V3.2)

DeepSeek
2025-12
Context
128K
Max Output
8K
Input $/1M
$0.28
Output $/1M
$0.42
MMLU
HumanEval

DeepSeek Reasoner (V3.2)

DeepSeek
2025-12
Context
128K
Max Output
64K
Input $/1M
$0.28
Output $/1M
$0.42
MMLU
HumanEval

Mistral Large 3

Mistral
2025-12
Context
262K
Max Output
33K
Input $/1M
$0.50
Output $/1M
$1.50
MMLU
HumanEval

Claude Haiku 4.5

Anthropic
2025-10
Context
200K
Max Output
64K
Input $/1M
$1.00
Output $/1M
$5.00
MMLU
HumanEval
88.1

GPT-5-mini

OpenAI
2025-09
Context
200K
Max Output
33K
Input $/1M
$0.25
Output $/1M
$2.00
MMLU
HumanEval

GPT-5-nano

OpenAI
2025-09
Context
200K
Max Output
33K
Input $/1M
$0.05
Output $/1M
$0.40
MMLU
HumanEval

Claude Sonnet 4.5

Anthropic
2025-09
Context
200K
Max Output
64K
Input $/1M
$3.00
Output $/1M
$15.00
MMLU
HumanEval
93.0

Gemini 2.5 Flash

Google
2025-09
Context
1.0M
Max Output
66K
Input $/1M
$0.30
Output $/1M
$2.50
MMLU
HumanEval

Grok 4

xAI
2025-09
Context
256K
Max Output
33K
Input $/1M
$3.00
Output $/1M
$15.00
MMLU
HumanEval

Grok 4 Fast

xAI
2025-09
Context
2.0M
Max Output
33K
Input $/1M
$0.20
Output $/1M
$0.50
MMLU
HumanEval

Mistral Medium 3.1

Mistral
2025-08
Context
128K
Max Output
33K
Input $/1M
$0.40
Output $/1M
$2.00
MMLU
HumanEval

o3

OpenAI
2025-06
Context
200K
Max Output
100K
Input $/1M
$2.00
Output $/1M
$8.00
MMLU
HumanEval

Mistral Small 3.2

Mistral
2025-06
Context
128K
Max Output
8K
Input $/1M
$0.10
Output $/1M
$0.30
MMLU
HumanEval

GPT-4.1

OpenAI
2025-04
Context
1.0M
Max Output
33K
Input $/1M
$2.00
Output $/1M
$8.00
MMLU
HumanEval

GPT-4.1-mini

OpenAI
2025-04
Context
1.0M
Max Output
33K
Input $/1M
$0.40
Output $/1M
$1.60
MMLU
HumanEval

GPT-4.1-nano

OpenAI
2025-04
Context
1.0M
Max Output
33K
Input $/1M
$0.10
Output $/1M
$0.40
MMLU
HumanEval

Llama 4 Maverick (Groq)

Meta/Groq
2025-04
Context
131K
Max Output
8K
Input $/1M
$0.20
Output $/1M
$0.60
MMLU
HumanEval

Llama 4 Scout (Groq)

Meta/Groq
2025-04
Context
131K
Max Output
8K
Input $/1M
$0.11
Output $/1M
$0.34
MMLU
HumanEval

Gemini 2.5 Pro

Google
2025-03
Context
1.0M
Max Output
66K
Input $/1M
$1.25
Output $/1M
$10.00
MMLU
HumanEval
Showing 28 of 41 models (13 legacy hidden)Last updated: April 12, 2026

What This Tool Does

AI Model Comparison Table is built for deterministic developer and agent workflows.

Compare 35+ AI models side by side: pricing, context windows, and specs across OpenAI GPT, Claude, Gemini, Grok, DeepSeek, Llama, and more.

Use How to Use for execution steps and FAQ for constraints, policies, and edge cases.

Last updated:

This tool is provided as-is for convenience. Output should be verified before use in any production or critical context.

Agent Invocation

Best Path For Builders

Browser workflow

Runs instantly in the browser with private local processing and copy/export-ready output.

Browser Workflow

This tool is optimized for instant in-browser execution with local data handling. Run it here and copy/export the output directly.

/ai-model-comparison/

For automation planning, fetch the canonical contract at /api/tool/ai-model-comparison.json.

How to Use AI Model Comparison Table

  1. 1

    Add models to compare

    Search for models by name (GPT, Claude, Gemini, Llama) or select from the curated list. You can add up to 10 models at once for comparison.

  2. 2

    View pricing details

    Check input/output token costs (per 1M tokens), see pricing tiers for different volume levels, and identify free tier models if cost is a priority.

  3. 3

    Compare specs side-by-side

    Filter by context window size, max output tokens, vision capabilities, and training data cutoff. Use these metrics to find the model that fits your use case.

  4. 4

    Check benchmarks and performance

    Review MMLU, GSM8K, and coding benchmark scores to see which model performs best on your specific task type (math, reasoning, code generation).

  5. 5

    Export and save your comparison

    Download the comparison as a CSV or PDF to share with your team or reference later when making model decisions.

Frequently Asked Questions

What is AI Model Comparison?
AI Model Comparison lets you compare AI models side by side across pricing, context windows, benchmarks, and technical specs. It's built for developers and teams evaluating which model fits their use case.
How do I use AI Model Comparison?
Select two or more AI models from the list, then compare their specs in a side-by-side table. You can filter by provider, sort by context window or benchmark scores, and view pricing breakdowns per million tokens.
Is AI Model Comparison free?
Yes. This tool is free to use with immediate access—no account required.
Does AI Model Comparison store or send my data?
No. All processing happens entirely in your browser. Your data never leaves your device — nothing is sent to any server.
How often is the model data updated?
Model data is refreshed automatically on a weekly basis to reflect the latest pricing changes, new model releases, and updated benchmark scores from providers like OpenAI, Anthropic, Google, and Meta.

AI Model Specifications — April 12, 2026

28 current models and 13 legacy models compared. Context windows, output limits, pricing, and release dates.

Current Models

Model Provider Context Max Output Input/1M Output/1M Released
GPT-5.4 OpenAI 1.0M 128K $2.50 $15.00 2026-03
GPT-5.4 Pro OpenAI 1.0M 128K $30.00 $180.00 2026-03
GPT-5.3-Codex OpenAI 400K 128K $1.75 $14.00 2026-02
GPT-5-mini OpenAI 200K 33K $0.25 $2.00 2025-09
GPT-5-nano OpenAI 200K 33K $0.050 $0.40 2025-09
o3 OpenAI 200K 100K $2.00 $8.00 2025-06
GPT-4.1 OpenAI 1.0M 33K $2.00 $8.00 2025-04
GPT-4.1-mini OpenAI 1.0M 33K $0.40 $1.60 2025-04
GPT-4.1-nano OpenAI 1.0M 33K $0.10 $0.40 2025-04
Claude Opus 4.6 Anthropic 200K 128K $5.00 $25.00 2026-02
Claude Sonnet 4.6 Anthropic 200K 64K $3.00 $15.00 2026-01
Claude Sonnet 4.5 Anthropic 200K 64K $3.00 $15.00 2025-09
Claude Haiku 4.5 Anthropic 200K 64K $1.00 $5.00 2025-10
Gemini 3.1 Pro Google 1.0M 66K $2.00 $12.00 2026-02
Gemini 3 Flash Google 1.0M 66K $0.50 $3.00 2026-01
Gemini 3.1 Flash-Lite Google 1.0M 66K $0.25 $1.50 2026-03
Gemini 2.5 Pro Google 1.0M 66K $1.25 $10.00 2025-03
Gemini 2.5 Flash Google 1.0M 66K $0.30 $2.50 2025-09
DeepSeek Chat (V3.2) DeepSeek 128K 8K $0.28 $0.42 2025-12
DeepSeek Reasoner (V3.2) DeepSeek 128K 64K $0.28 $0.42 2025-12
Mistral Large 3 Mistral 262K 33K $0.50 $1.50 2025-12
Mistral Medium 3.1 Mistral 128K 33K $0.40 $2.00 2025-08
Mistral Small 3.2 Mistral 128K 8K $0.10 $0.30 2025-06
Llama 4 Maverick (Groq) Meta/Groq 131K 8K $0.20 $0.60 2025-04
Llama 4 Scout (Groq) Meta/Groq 131K 8K $0.11 $0.34 2025-04
Grok 4 xAI 256K 33K $3.00 $15.00 2025-09
Grok 4.1 Fast xAI 2.0M 33K $0.20 $0.50 2026-01
Grok 4 Fast xAI 2.0M 33K $0.20 $0.50 2025-09
Legacy Models (13) — still available in APIs
Model Provider Context Input/1M Output/1M Released
GPT-5.2 OpenAI 400K $1.75 $14.00 2026-01
GPT-5.1 OpenAI 400K $1.25 $10.00 2025-10
GPT-5 OpenAI 400K $1.25 $10.00 2025-08
o3-pro OpenAI 200K $20.00 $80.00 2025-07
o4-mini OpenAI 200K $1.10 $4.40 2025-07
o3-mini OpenAI 200K $1.10 $4.40 2025-01
GPT-4o OpenAI 128K $2.50 $10.00 2024-05
GPT-4o-mini OpenAI 128K $0.15 $0.60 2024-07
Claude Opus 4.5 Anthropic 200K $5.00 $25.00 2025-11
Claude Sonnet 4 Anthropic 200K $3.00 $15.00 2025-05
Gemini 2.0 Flash Google 1.0M $0.10 $0.40 2025-02
Mistral Large 2 Mistral 128K $2.00 $6.00 2024-07
Llama 3.3 70B Meta/Groq 131K $0.59 $0.79 2024-12

Market Trends — April 12, 2026

  • Competition across OpenAI, Anthropic, Google, xAI, DeepSeek, and open-weight ecosystems continues to compress pricing.
  • High-context tiers are becoming standard across major providers, improving long-document and agent workflows.
  • Fast and budget model tiers are increasingly viable for high-volume production use cases.
  • Prompt caching and retrieval-aware workflows are now key levers for reducing effective per-request cost.
  • Model selection is moving from single-model strategies to portfolio-based routing by task, latency, and budget.

Key Comparisons

OpenAI GPT vs Claude Flagships — Compare pricing, max output, and context windows for top-tier reasoning and coding performance.

Claude Balanced Tier vs GPT Balanced Tier — Useful when you need strong quality without paying flagship rates.

Gemini vs GPT Families — Often strong on context capacity and multimodal workflows; compare against your latency and cost targets.

DeepSeek vs Open-Weight Llama/Mistral Options — Good comparison set for budget-sensitive production pipelines.

Fast/Low-Cost Tiers Across Providers — Compare context limits and output pricing before committing to high-volume jobs.