AI Agents 相关度: 9/10

Mitigating Conversational Inertia in Multi-Turn Agents

Yang Wan, Zheng Cao, Zhenhao Zhang, Zhengwen Zeng, Shuheng Shen, Changhua Meng, Linchao Zhu
arXiv: 2602.03664v1 发布: 2026-02-03 更新: 2026-02-03

AI 摘要

该论文研究了多轮Agent中的对话惯性问题,并提出通过上下文偏好学习降低惯性,提升性能。

主要贡献

  • 发现了LLM Agent中的对话惯性现象
  • 提出了基于上下文偏好学习的解决方法
  • 提出了平衡探索和利用的上下文管理策略

方法论

通过注意力分析识别对话惯性,构建偏好对进行学习,并设计上下文管理策略。

原文摘要

Large language models excel as few-shot learners when provided with appropriate demonstrations, yet this strength becomes problematic in multiturn agent scenarios, where LLMs erroneously mimic their own previous responses as few-shot examples. Through attention analysis, we identify conversational inertia, a phenomenon where models exhibit strong diagonal attention to previous responses, which is associated with imitation bias that constrains exploration. This reveals a tension when transforming few-shot LLMs into agents: longer context enriches environmental feedback for exploitation, yet also amplifies conversational inertia that undermines exploration. Our key insight is that for identical states, actions generated with longer contexts exhibit stronger inertia than those with shorter contexts, enabling construction of preference pairs without environment rewards. Based on this, we propose Context Preference Learning to calibrate model preferences to favor low-inertia responses over highinertia ones. We further provide context management strategies at inference time to balance exploration and exploitation. Experimental results across eight agentic environments and one deep research scenario validate that our framework reduces conversational inertia and achieves performance improvements.

标签

LLM Agent 对话惯性 偏好学习 上下文管理

arXiv 分类

cs.AI cs.LG