LLM Comparison Table
Compare 42 large language models across pricing Cost per 1 million tokens for input and output. Input tokens are what you send; output tokens are what the model generates. , context window The maximum number of tokens (roughly ¾ of a word each) a model can process in a single request — including your prompt, conversation history, and the model's response. , speed Tokens generated per second. Higher means faster responses. Varies by provider load and prompt length. , and quality benchmarks Standardised tests like MMLU (knowledge), GPQA (reasoning), HumanEval (coding), and Chatbot Arena (human preference). . Click any column header to sort.
Last updated: 14 Apr 2026 · Prices from official API documentation
Head-to-Head Compare42 models
| ModelModel | ProviderProvider | Input /1MIn $/1M | Output /1MOut $/1M | ContextCtx | SpeedSpeed | QualityQual. | ValueValue |
|---|---|---|---|---|---|---|---|
| Mistral | $0.02 | $0.04 | 131K | 0 t/s | 72 | 20571 | |
| Amazon | $0.04 | $0.14 | 128K | 0 t/s | 68 | 5978 | |
| Cohere | $0.04 | $0.15 | 128K | 0 t/s | 65 | 5328 | |
| Reka | $0.10 | $0.20 | 66K | 0 t/s | 74 | 4229 | |
| Amazon | $0.06 | $0.24 | 300K | 0 t/s | 72 | 3692 | |
| $0.07 | $0.30 | 1.0M | 450 t/s | 76 | 3118 | ||
| Meta | $0.10 | $0.32 | 131K | 80 t/s | 79 | 2981 | |
| OpenAI | $0.05 | $0.40 | 400K | 0 t/s | 78 | 2496 | |
| $0.10 | $0.40 | 1.0M | 400 t/s | 81 | 2492 | ||
| $0.10 | $0.40 | 1.0M | 0 t/s | 78 | 2400 | ||
| OpenAI | $0.10 | $0.40 | 1.0M | 200 t/s | 75 | 2308 | |
| Alibaba | $0.12 | $0.39 | 33K | 65 t/s | 71 | 2202 | |
QwQ 32BOSS | Alibaba | $0.15 | $0.58 | 131K | 0 t/s | 78 | 1651 |
| OpenAI | $0.15 | $0.60 | 128K | 150 t/s | 80 | 1641 | |
| Meta | $0.15 | $0.60 | 1.0M | 95 t/s | 75 | 1538 | |
| Mistral | $0.35 | $0.56 | 128K | 150 t/s | 76 | 1498 | |
| Cohere | $0.15 | $0.60 | 128K | 0 t/s | 73 | 1497 | |
| DeepSeek | $0.20 | $0.77 | 164K | 49 t/s | 86 | 1371 | |
DeepSeek V3OSS | DeepSeek | $0.32 | $0.89 | 164K | 0 t/s | 76 | 1017 |
| Perplexity | $1.00 | $1.00 | 127K | 0 t/s | 74 | 740 | |
| Alibaba | $0.46 | $1.82 | 131K | 40 t/s | 87 | 588 | |
DeepSeek R1OSS | DeepSeek | $0.70 | $2.50 | 64K | 30 t/s | 85 | 415 |
| $0.30 | $2.50 | 1.0M | 350 t/s | 78 | 400 | ||
| Amazon | $0.80 | $3.20 | 300K | 0 t/s | 78 | 300 | |
| Anthropic | $0.80 | $4.00 | 200K | 120 t/s | 82 | 256 | |
| OpenAI | $1.10 | $4.40 | 200K | 65 t/s | 90 | 252 | |
| Mistral | $2.00 | $6.00 | 128K | 80 t/s | 86 | 172 | |
| OpenAI | $2.00 | $8.00 | 200K | 15 t/s | 88 | 135 | |
| OpenAI | $2.00 | $8.00 | 1.0M | 110 t/s | 77 | 118 | |
| OpenAI | $1.25 | $10.00 | 400K | 75 t/s | 87 | 111 | |
| $1.25 | $10.00 | 1.0M | 90 t/s | 83 | 106 | ||
Command AOSS | Cohere | $2.50 | $10.00 | 256K | 0 t/s | 82 | 101 |
| Cohere | $2.50 | $10.00 | 128K | 0 t/s | 79 | 97 | |
| OpenAI | $1.75 | $14.00 | 400K | 85 t/s | 90 | 82 | |
| xAI | $3.00 | $15.00 | 256K | 50 t/s | 88 | 73 | |
| Anthropic | $3.00 | $15.00 | 1.0M | 90 t/s | 86 | 72 | |
| xAI | $3.00 | $15.00 | 131K | 70 t/s | 85 | 71 | |
| Anthropic | $3.00 | $15.00 | 1.0M | 80 t/s | 79 | 66 | |
| OpenAI | $6.00 | $18.00 | 128K | 100 t/s | 75 | 50 | |
| Anthropic | $15.00 | $75.00 | 1.0M | 50 t/s | 89 | 15 | |
| OpenAI | $20.00 | $80.00 | 200K | 0 t/s | 88 | 14 | |
| Anthropic | $15.00 | $75.00 | 200K | 30 t/s | 84 | 14 |
Prices per 1M tokens (USD)Speed in output tokens/secondQuality: composite benchmark score (0-100)Value: quality per unit cost