I Need More Context

Model Max Context Window Notes
Gemini 1.5 Pro1 million tokensEarly access model from Google. Incredible length support.
Claude 2.1200,000 tokensExtremely large context window. Very strong at long-document tasks.
Claude 2.0100,000 tokensPrevious version, still widely used.
Command R+ (Cohere)128,000 tokensRAG-optimized, good for enterprise use.
GPT-4 Turbo128,000 tokensCurrent ChatGPT model (fast + cheap). Used in ChatGPT and OpenAI API.
GPT-432,768 tokensEarlier GPT-4 versions.
Gemini 1.0 Pro~32,000 tokensCompetitive with GPT-4-32k.
Mistral (Mixtral)~32,000 tokensOpen-weight MoE model, efficient. Context limit estimated.
GPT-3.5 Turbo16,385 tokensAvailable via API.
Mistral 7B8,192 tokensSmaller open-weight model.
GPT-3.54,096 tokensDefault, older GPT-3.5 model.
LLaMA 2 70B4,096 tokensPopular open-weight base model.
LLaMA 3 (expected)TBDExpected to support larger windows.
Get updates when companies make memory related progress/features:
Join builders working on memory related AI problems: