Context Window

Which LLM context windowfits your use case?

Compare context windows, rank models by use case, and calculate how much Mem0 can reduce your token usage.

Models1051+
Largest context10M

What you get

Every LLM context window, ranked and ready

Browse every model by how much text it can handle, find the right one for your use case, and see how Mem0 reduces what you need to send.

Models indexed

1051+

Providers covered

37

Largest context

10M

Use-case rankings

8

Context window spectrum

Models grouped by how much text they can handle

4K32K128K1M10M
4K196 models
32K322 models
128K468 models
1M54 models
10M4 models

Largest: Llama 4 Scout 17b 128e Instruct Maas at 10M tokens

Rankings by use case

The right model depends on more than size. See what actually works for each job.

All 8 rankings

Mem0 cuts context by ~80%

Mem0 remembers what matters so you don't have to send the full history every time. Smaller requests means faster, cheaper AI.

Without Mem0100K tokens
With Mem0~20K tokens
80% less context Faster responses Fewer hallucinations
Calculate your savings

All Models

1051 models

Llama 4 Scout 17b 128e Instruct MaasMeta
10M
Llama 4 Scout 17b 16eMeta
10M
Llama 4 Scout 17b 16e Instruct Fp8Meta
10M
Llama 4 Scout 17b 16e Instruct MaasMeta
10M
Gemini Exp 1206Google
2.1M
Grok 4 1 FastXai
2M
Grok 4 1 Fast Non Reasoning LatestXai
2M
Grok 4 1 Fast Reasoning LatestXai
2M

Showing 8 of 1051. View all →

Rankings

Models ranked by use case

Not all models are equal for every job. Pick the ranking that matches what you're building.

View all rankings

Largest Context Window

1051

Models ranked by maximum context window size.

#1Llama 4 Scout 17b 128e Instruct Maas10M
Llama 4 Scout 17b 128e Instruct Maas
10M
Llama 4 Scout 17b 16e
10M
Llama 4 Scout 17b 16e Instruct Fp8
10M
View ranking

Best for RAG

1051

Models best suited for Retrieval-Augmented Generation workloads.

#1Llama 4 Scout 17b 128e Instruct Maas10M
Llama 4 Scout 17b 128e Instruct Maas
10M
Llama 4 Scout 17b 16e
10M
Llama 4 Scout 17b 16e Instruct Fp8
10M
View ranking

Best for AI Agents

1051

Models with the capabilities needed to power autonomous agent workflows.

#1Llama 4 Scout 17b 128e Instruct Maas10M
Llama 4 Scout 17b 128e Instruct Maas
10M
Llama 4 Scout 17b 16e
10M
Llama 4 Scout 17b 16e Instruct Fp8
10M
View ranking

Best for Document Processing

1051

Models with large enough context to process long documents in one pass.

#1Llama 4 Scout 17b 128e Instruct Maas10M
Llama 4 Scout 17b 128e Instruct Maas
10M
Llama 4 Scout 17b 16e
10M
Llama 4 Scout 17b 16e Instruct Fp8
10M
View ranking

Best Value per Context Token

1051

Models offering the most context window per dollar of input cost.

#1Llama 4 Scout 17b 16e10M
Llama 4 Scout 17b 16e
10M
Llama 4 Scout 17b 128e Instruct Maas
10M
Llama 4 Scout 17b 16e Instruct Maas
10M
View ranking

Best Multimodal

1051

Vision-capable models with the largest context windows.

#1Llama 4 Scout 17b 16e10M
Llama 4 Scout 17b 16e
10M
Gemini Exp 1206
2.1M
Grok 4 1 Fast
2M
View ranking

Best Reasoning

1051

Models with extended thinking or strong reasoning capabilities.

#1Grok 4 1 Fast2M
Grok 4 1 Fast
2M
Grok 4 1 Fast Reasoning Latest
2M
Grok 4.20
2M
View ranking

Best for Chatbots

1051

Fast, cost-efficient models ideal for real-time conversational applications.

#1Llama 4 Scout 17b 128e Instruct Maas10M
Llama 4 Scout 17b 128e Instruct Maas
10M
Llama 4 Scout 17b 16e
10M
Llama 4 Scout 17b 16e Instruct Fp8
10M
View ranking

Powered by Mem0

Use a smaller model.
Get better results.

Mem0 gives your AI long-term memory so you stop re-sending context on every call. That means you can use a smaller, faster, cheaper model — and still get better answers.

Example: a multi-turn chat session

Without Mem0~128K tokens sent
Full history
Repeated info
Old context
With Mem0~20K tokens sent
Key memories
Current turn

80% less to send — works with any model