All AI chatbots have the ability to understand the context of a conversation. However, there is only so much these models can “remember” and this is known as the “Context Length” or “Context Window.”
In other words, each AI Model can only process so much of your conversation before it starts to “forget” things.
For example, OpenAI’s GPT-4o model has a context length of 128,000 tokens, which is the equivalent of around 90,000 words. It can only actively process the last 90,000 words of a conversation.
If a conversation extends beyond this 90,000 word context length, GPT-4o will start “forgetting” the earlier parts of the exchange and may produce irrelevant or confusing responses.
Here is a breakdown of each of Magai’s available AI Models and their context length:
Model Name | Context Length |
Auto | 128K |
Claude 3.5 Haiku | 200K |
Claude 3.5 Sonnet | 200K |
Claude 3.7 Sonnet | 200K |
DeepSeek R1 | 64K |
DeepSeek V3 | 64K |
Gemini 2.0 Flash | 1M |
Gemini 2.0 Pro | 2M |
Gemini 2.0 Thinking | 1M |
GPT-4.5 | 128K |
GPT-4o | 128K |
GPT-4o Mini | 128K |
Grok 2 | 131K |
Grok 2 Vision | 33K |
Llama 3.1 405B | 131K |
Llama 3.2 11B | 130K |
Llama 3.3 70B | 128K |
Mistral Large 2411 | 128K |
Nemotron 70B | 131K |
Nova Lite | 300K |
Nova Micro | 128K |
Nova Pro | 300K |
o1 | 200K |
o1 Mini | 128K |
o3 Mini | 200K |
Perplexity Sonar Online | 127K |
Pixtral Large 2411 | 128K |
Most AI assistants do not notify the user when the context length is exceeded, which can lead to a poor experience.
To address this, our team was careful to craft a subtle and transparent way of indicating where a conversations context is cut off.
We recommend using our Claude AI model which has a context length of 200,000 words to avoid these issues with reaching the AI’s context length limit and having more in-depth conversations.