/KON-tekst WIN-doh/
The maximum amount of text an AI can 'see' at once — including both your prompt and its response. Measured in tokens.
The context window is the AI's working memory. It's the total number of tokens the model can process in a single conversation — everything you've said, everything it's said back, and everything in the system prompt. When you exceed it, the AI starts 'forgetting' earlier parts of the conversation.
Claude offers 200K tokens (~150K words). GPT-4 offers 128K tokens. These sound huge, but they fill up fast when you're pasting codebases, documents, or long conversation histories. Understanding context windows explains why AI conversations 'drift' after a while — the model literally can't see what you discussed earlier.
Practical context window management is a core AI operator skill: knowing when to summarize, when to start a new conversation, and how to structure prompts to use context efficiently.
When planning prompts, debugging conversation drift, or deciding how much context to include in an AI interaction.
Context window management is the difference between AI that remembers and AI that forgets. Master it and your AI interactions become dramatically more effective.
It's literally a window — you can only see what fits in the frame. Scroll too far and the beginning disappears.
A Mac app that coaches your AI vocabulary daily