L LLM Cloud Hub
Glossary

Context window

The maximum number of tokens an LLM can process in a single request.

The maximum number of tokens (input + output) a model can process in a single request. A 128k context window can fit ~96k words — about a 350-page book. Larger windows enable RAG, long-document summarization, and full-codebase reasoning.

See also
← Back to full glossary

Keyboard shortcuts

?
Show this overlay
/
Focus the first form field
g h
Go to / (home)
g b
Go to /best-llm-for
g c
Go to /cost
g s
Go to /self-hosted
g x
Go to /compliance
Esc
Close any overlay

Inspired by Linear and GitHub conventions. The two-key sequences (g then h) work within ~1 second.