Information supplied to a language model at query time through retrieval-augmented generation — distinct from parametric knowledge encoded in the model’s weights.
Retrieval knowledge is information supplied to a language model at query time through retrieval-augmented generation — documents or passages fetched from an external index and inserted into the model’s context window before it generates a response. It is distinct from parametric knowledge, which is encoded in the model’s weights during training.
Retrieval knowledge is the layer that content and technical optimization directly influences. When a query triggers retrieval, the model answers based on what it found — which means the content that gets retrieved and passes reranking shapes the response. For brands with absent or inaccurate parametric representations, retrieval knowledge is the faster correction path: publish citation-ready content, ensure it indexes, and it can appear in AI responses before the next training run happens.
Retrieval knowledge overrides parametric knowledge reliably. It does not — when parametric inertia is high, the model may favor its encoded belief over conflicting retrieved content. Retrieval is the faster lever, but not always the stronger one.