Guide
What is Context Window?
Understanding LLM token limits and their implications.
The context window is the maximum number of tokens an LLM can process in a single request. It includes your prompt, system instructions, and the model's response.
Context Window Sizes
- GPT-4: 8K-128K tokens
- Claude: Up to 200K tokens
- Gemini: Up to 1M tokens
- Open-source models: Varies widely
Context Window Considerations
- Larger context = higher cost
- Performance may degrade with very long contexts
- "Lost in the middle" problem
- RAG can work around limits
How many tokens is 1 word?
Roughly 1.3 tokens per word in English. A 4K token context is about 3,000 words.
Track token usage
Start Free