Largest Context Window
Models ranked by maximum context window size in tokens. Larger context windows allow processing more text in a single request.
Methodology: Sorted by context_window_tokens descending. All published models with context window data included.
| # | Model | Provider | Metric |
|---|---|---|---|
| 1 | Gemini 1.5 Pro | Google DeepMind | 2,000,000 tokens |
| 2 | Gemini 1.5 Flash | Google DeepMind | 1,000,000 tokens |
| 3 | Gemini 2.0 Flash | Google DeepMind | 1,000,000 tokens |
| 4 | o1 | OpenAI | 200,000 tokens |
| 5 | Claude 3 Opus | Anthropic | 200,000 tokens |
| 6 | Claude 3 Sonnet | Anthropic | 200,000 tokens |
| 7 | Claude 3 Haiku | Anthropic | 200,000 tokens |
| 8 | Claude 3.5 Sonnet | Anthropic | 200,000 tokens |
| 9 | Claude 3.5 Haiku | Anthropic | 200,000 tokens |
| 10 | Claude Sonnet 4.6 | Anthropic | 200,000 tokens |
| 11 | Grok 2 | xAI | 131,072 tokens |
| 12 | GPT-4 Turbo | OpenAI | 128,000 tokens |
| 13 | GPT-4o | OpenAI | 128,000 tokens |
| 14 | GPT-4o mini | OpenAI | 128,000 tokens |
| 15 | DeepSeek V3 | DeepSeek | 128,000 tokens |
| 16 | Llama 3.1 405B | Meta | 128,000 tokens |
| 17 | Llama 3.3 70B | Meta | 128,000 tokens |
| 18 | Mistral Large 2 | Mistral AI | 128,000 tokens |
| 19 | Command R+ | Cohere | 128,000 tokens |
| 20 | Qwen 2.5 72B | Alibaba Cloud (Qwen) | 128,000 tokens |
| 21 | Claude 2 | Anthropic | 100,000 tokens |
| 22 | DeepSeek R1 | DeepSeek | 64,000 tokens |
| 23 | Gemini 1.0 Pro | Google DeepMind | 32,768 tokens |
| 24 | Mistral 7B | Mistral AI | 32,768 tokens |
| 25 | Mixtral 8x7B | Mistral AI | 32,768 tokens |
| 26 | GPT-3.5 Turbo | OpenAI | 16,385 tokens |
| 27 | GPT-4 | OpenAI | 8,192 tokens |
| 28 | Llama 3 70B | Meta | 8,192 tokens |
| 29 | GPT-3 (davinci-002) | OpenAI | 4,096 tokens |
| 30 | Llama 2 70B | Meta | 4,096 tokens |