AI Agents 相关度: 9/10

TKG-Thinker: Towards Dynamic Reasoning over Temporal Knowledge Graphs via Agentic Reinforcement Learning

Zihao Jiang, Miao Peng, Zhenyan Shan, Wenjie Xu, Ben Liu, Gong Chen, Ziqi Gao, Min Peng
arXiv: 2602.05818v1 发布: 2026-02-05 更新: 2026-02-05

AI 摘要

TKG-Thinker通过Agentic强化学习进行时序知识图谱动态推理,提升复杂时序约束下的推理能力。

主要贡献

  • 提出了TKG-Thinker,一个用于时序知识图谱推理的智能体。
  • 使用双重训练策略(SFT+RL)提高智能体的规划和推理能力。
  • 在benchmark数据集上取得了SOTA性能,并展现出良好的泛化能力。

方法论

使用监督微调(SFT)赋予智能体规划能力,再通过强化学习(RL)优化推理策略,实现动态交互推理。

原文摘要

Temporal knowledge graph question answering (TKGQA) aims to answer time-sensitive questions by leveraging temporal knowledge bases. While Large Language Models (LLMs) demonstrate significant potential in TKGQA, current prompting strategies constrain their efficacy in two primary ways. First, they are prone to reasoning hallucinations under complex temporal constraints. Second, static prompting limits model autonomy and generalization, as it lack optimization through dynamic interaction with temporal knowledge graphs (TKGs) environments. To address these limitations, we propose \textbf{TKG-Thinker}, a novel agent equipped with autonomous planning and adaptive retrieval capabilities for reasoning over TKGs. Specifically, TKG-Thinker performs in-depth temporal reasoning through dynamic multi-turn interactions with TKGs via a dual-training strategy. We first apply Supervised Fine-Tuning (SFT) with chain-of thought data to instill core planning capabilities, followed by a Reinforcement Learning (RL) stage that leverages multi-dimensional rewards to refine reasoning policies under intricate temporal constraints. Experimental results on benchmark datasets with three open-source LLMs show that TKG-Thinker achieves state-of-the-art performance and exhibits strong generalization across complex TKGQA settings.

标签

Temporal Knowledge Graph Reinforcement Learning Agent Reasoning

arXiv 分类

cs.AI cs.DB