LLM Reasoning 相关度: 9/10

RADAR: Reasoning as Discrimination with Aligned Representations for LLM-based Knowledge Graph Reasoning

Bo Xue, Yuan Jin, Luoyi Fu, Jiaxin Ding, Xinbing Wang
arXiv: 2602.21951v1 发布: 2026-02-25 更新: 2026-02-25

AI 摘要

RADAR通过判别式学习提升LLM在知识图谱推理中的泛化能力和鲁棒性。

主要贡献

  • 提出RADAR框架,将知识图谱推理重构为判别式实体选择任务
  • 利用强化学习增强实体可分离性,优化表示空间
  • 实验证明RADAR在链接预测和三元组分类上优于现有LLM基线

方法论

RADAR采用判别式学习,通过强化学习优化实体表示的可分离性,并在表示空间进行推理。

原文摘要

Knowledge graph reasoning (KGR) infers missing facts, with recent advances increasingly harnessing the semantic priors and reasoning abilities of Large Language Models (LLMs). However, prevailing generative paradigms are prone to memorizing surface-level co-occurrences rather than learning genuine relational semantics, limiting out-of-distribution generalization. To address this, we propose RADAR, which reformulates KGR from generative pattern matching to discriminative relational reasoning. We recast KGR as discriminative entity selection, where reinforcement learning enforces relative entity separability beyond token-likelihood imitation. Leveraging this separability, inference operates directly in representation space, ensuring consistency with the discriminative optimization and bypassing generation-induced hallucinations. Across four benchmarks, RADAR achieves 5-6% relative gains on link prediction and triple classification over strong LLM baselines, while increasing task-relevant mutual information in intermediate representations by 62.9%, indicating more robust and transferable relational reasoning.

标签

知识图谱推理 大型语言模型 判别式学习 强化学习

arXiv 分类

cs.CL