Reification Gap in LLM Modeling
The gap between what language models manipulate (statistical patterns in text) and what they appear to represent (concepts, facts, reasoning). LLMs work with token sequences, not the things those tokens refer to.
The Gap
When we ask an LLM about “Paris,” it doesn’t access some internal representation of Paris-the-city. It predicts tokens based on patterns of how “Paris” co-occurs with other tokens in training data.
Implications
Brittle Grounding: Model “knowledge” is surface patterns, not grounded understanding Sensitivity to Framing: Same question phrased differently gets different answers Hallucination Origin: Models generate plausible-sounding sequences without verifying reference Limits of Scaling: More data improves pattern coverage, not referential grounding
Why It Matters
The reification gap explains why LLMs can be simultaneously impressive (fluent, knowledgeable-seeming) and unreliable (confident confabulation). They’re doing something different from what their outputs suggest.
Mitigation
- Ground LLM outputs in external knowledge sources (RAG)
- Use verification steps before consequential actions
- Don’t treat LLM outputs as authoritative references
Related: 05-atom—uniform-confidence-problem, 05-atom—hallucination-inherent