What is an LLM Context Window?
An LLM context window comparison chart directly visualizes how much text, measured in tokens, a large language model can process and retain simultaneously. This limit is crucial because it defines the AI’s short-term memory for any given interaction, directly influencing its ability to recall past information and maintain coherence.
Defining the LLM Context Window
The LLM context window represents the maximum sequence length, in tokens, that a model can ingest and consider when generating a response. It dictates how much prior conversation or document content the AI “remembers” for its immediate task. This constraint is fundamental to understanding an LLM’s capabilities and limitations.
Understanding Tokenization
LLMs process text by first breaking it down into tokens. A token is a common sequence of characters found in text, which can be a word, part of a word, or punctuation. For English text, approximately 100 tokens equate to about 75 words. This conversion is vital for estimating the actual amount of human-readable text that fits within a model’s context window.