<RETURN_TO_BASE

Graph-R1: Agentic Hypergraph RAG for Multi-Turn Reinforced Reasoning

'Graph-R1 combines hypergraph knowledge, agentic multi-turn retrieval, and end-to-end RL to deliver state-of-the-art QA accuracy and efficient generation.'

What Graph-R1 brings

Graph-R1 is an agentic GraphRAG framework that combines hypergraph knowledge representation, multi-turn agentic retrieval, and end-to-end reinforcement learning. It targets the common RAG weakness of chunk-based retrieval and one-shot reasoning by encoding richer n-ary relationships and by letting a retrieval agent iteratively explore, refine, and produce answers.

Lightweight hypergraph knowledge construction

Instead of indexing text chunks, Graph-R1 builds a knowledge hypergraph where each segment is extracted using LLM-driven n-ary relation extraction. This produces semantically richer connections between entities while keeping cost and computation manageable. Reported construction figures show 5.69 seconds and $2.81 per 1,000 tokens for building the hypergraph, producing graphs with about 120,499 nodes and 98,073 edges.

Agentic multi-turn retrieval loop

Retrieval is modeled as a multi-turn interaction loop: think → retrieve → rethink → generate. At each step the agent decides whether to continue exploring the hypergraph or to stop and answer. Graph-R1 fuses entity-based and direct hyperedge retrieval via reciprocal rank aggregation, enabling adaptive focus on high-impact graph regions. Typical interactions average 2.3–2.5 turns and use concise context exchanges of roughly 1,200–1,500 tokens.

End-to-end RL with GRPO

Graph-R1 optimizes its agents using Group Relative Policy Optimization (GRPO). The reward function combines format rewards (ensuring structural coherence of the reasoning trajectory) and answer rewards (semantic accuracy). Only answers produced along structurally valid reasoning paths receive full reward, which encourages agents to develop reliable, generalizable reasoning policies tied to the hypergraph structure.

Benchmark performance

Graph-R1 was evaluated on six QA datasets including 2WikiMultiHopQA, HotpotQA, Musique, Natural Questions, PopQA, and TriviaQA. Using Qwen2.5-7B as a backbone, reported average F1 scores are:

  • NaiveGeneration: 13.87
  • StandardRAG: 15.89
  • GraphRAG: 24.87
  • HyperGraphRAG: 29.40
  • Search-R1: 46.19
  • R1-Searcher: 42.29
  • Graph-R1: 57.82

Graph-R1 achieves up to 57.82 average F1, substantially outperforming prior graph-based and RAG baselines, with larger base models further amplifying gains.

Ablations and robustness

Ablation studies show that removing any of the main modules—hypergraph construction, multi-turn reasoning, or RL optimization—dramatically drops performance. Cross-validation on out-of-distribution settings demonstrates strong generalizability, with O.O.D./I.I.D. ratios frequently above 85%.

Efficiency, generation cost, and quality

Despite richer structure, Graph-R1 remains efficient: response time is reported at about 7.0 seconds per query and per-query generation cost is effectively $0, outperforming competitors like HyperGraphRAG which reports 9.6 seconds and $8.76 per query. Generation quality was also evaluated across seven dimensions, and Graph-R1 scored highest in correctness (86.9), relevance (95.2), and coherence (88.5).

Theoretical insights and applicability

Information-theoretic analysis suggests graph-structured knowledge yields higher information density per retrieval and faster convergence to correct answers compared to chunk-based methods. Multi-turn interaction improves retrieval efficiency by focusing on high-impact regions, and end-to-end RL helps align structured evidence with natural language outputs.

Graph-R1 is especially relevant for knowledge-intensive domains that demand accuracy and interpretability, such as healthcare, law, and enterprise knowledge automation. The framework charts a path toward more agentic, transparent, and reliable retrieval-augmented systems.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский