L LLM Cloud Hub
Glossary

RAG (Retrieval-Augmented Generation)

Fetching relevant documents and prepending them to the prompt for grounded answers.

A pattern where you fetch relevant documents from your own corpus (via vector search or keyword), then prepend them to the prompt. The LLM is grounded in your data without retraining. Workloads are heavy on input tokens.

See also
← Back to full glossary

Keyboard shortcuts

?
Show this overlay
/
Focus the first form field
g h
Go to / (home)
g b
Go to /best-llm-for
g c
Go to /cost
g s
Go to /self-hosted
g x
Go to /compliance
Esc
Close any overlay

Inspired by Linear and GitHub conventions. The two-key sequences (g then h) work within ~1 second.