Multi-Hop Reasoning Failure

The Problem

Queries requiring chaining multiple pieces of information fail because LLM cannot connect facts across separate chunks or perform multi-step inference.

Symptoms

  • ❌ Can't answer questions needing 2+ facts combined

  • ❌ "I don't have that information" despite facts present

  • ❌ Answers partial chain, misses full connection

  • ❌ Cannot infer transitive relationships

  • ❌ Fails on "who has access to X?" type queries

Real-World Example

Knowledge base contains (separate docs):
→ Doc A: "Alice is member of Engineering team"
→ Doc B: "Engineering team has access to Production DB"

Query: "Does Alice have access to Production DB?"

Required reasoning:
1. Alice → Engineering team (from Doc A)
2. Engineering team → Production DB (from Doc B)
3. Therefore: Alice → Production DB ✓

AI response: "I don't have information about Alice's access to Production DB."

Failed to chain: Both facts retrieved but not connected

Deep Technical Analysis

Single-Hop vs Multi-Hop

Single-Hop (Easy):

Multi-Hop (Hard):

Retrieval Limitations

Fact Dispersion:

Missing Intermediate:

Prompting for Reasoning

Chain-of-Thought:

Structured Extraction:

Knowledge Graph Approach

Graph Structure:


How to Solve

Use chain-of-thought prompting for multi-step queries + implement two-stage: extract facts, then reason + build knowledge graph for entities/relationships + ensure all related chunks retrieved (expand retrieval for graph queries) + use higher-capability models (GPT-4 over GPT-3.5) for reasoning + test multi-hop eval set to measure success rate. See Multi-Hop Reasoning.

Last updated