LLM Hallucination in Knowledge Engineering
The specific challenge of LLM confabulation when used for knowledge engineering tasks, creating, extending, or validating knowledge structures.
Manifestations
Relationship Invention: Creating plausible but false relationships between entities Property Confabulation: Assigning properties entities don’t have Hierarchy Errors: Placing concepts in wrong taxonomic positions Citation Fabrication: Generating nonexistent sources Definition Drift: Subtly redefining terms from their intended meaning
Why Knowledge Engineering is Vulnerable
- Requires factual accuracy (hallucination is directly harmful)
- Involves specialized domains (may be underrepresented in training)
- Errors propagate through connected knowledge structures
- Human reviewers may lack expertise to catch subtle errors
Mitigation Strategies
Verification Pipeline:
- Generate with LLM
- Verify against authoritative sources
- Human expert review for novel claims
Prompt Engineering:
- Request only high-confidence assertions
- Ask for source citations (then verify them)
- Use conservative extraction prompts
Hybrid Approaches:
- LLM for candidate generation
- Rule-based validation
- Human final approval
Related: 05-atom—uniform-confidence-problem, 06-molecule—knowledge-graph-construction