Zum Inhalt springen
Agent Hub
Back to glossary
models

Context window

The context window is the maximum amount of text (measured in tokens) a language model can process at once — typically 128k to 1M tokens with current models.

Also known as: Kontext-Fenster

In detail

The window includes system prompt, conversation history, retrieved RAG documents, and the current request. Bigger window = more context for better answers, but also higher cost and latency.

Rule of thumb: 1 token ≈ 0.75 English words. A 128k window holds roughly a 200-page book.

Related terms