LLM Reasoning 相关度: 8/10

Beyond the Assistant Turn: User Turn Generation as a Probe of Interaction Awareness in Language Models

Sarath Shekkizhar, Romain Cosentino, Adam Earle
arXiv: 2604.02315v1 发布: 2026-04-02 更新: 2026-04-02

AI 摘要

提出用户回复生成作为探测LLM交互意识的方法,发现交互意识与任务准确率解耦,可通过后训练提升。

主要贡献

  • 提出一种新的评估LLM交互意识的probe:用户回复生成。
  • 揭示了LLM的交互意识与任务准确率之间存在解耦现象。
  • 展示了通过 collaboration-oriented post-training 可以提高交互意识。

方法论

通过让LLM在给定对话历史后生成用户回复,评估其生成的回复是否与上下文相关,并进行扰动实验验证。

原文摘要

Standard LLM benchmarks evaluate the assistant turn: the model generates a response to an input, a verifier scores correctness, and the analysis ends. This paradigm leaves unmeasured whether the LLM encodes any awareness of what follows the assistant response. We propose user-turn generation as a probe of this gap: given a conversation context of user query and assistant response, we let a model generate under the user role. If the model's weights encode interaction awareness, the generated user turn will be a grounded follow-up that reacts to the preceding context. Through experiments across $11$ open-weight LLMs (Qwen3.5, gpt-oss, GLM) and $5$ datasets (math reasoning, instruction following, conversation), we show that interaction awareness is decoupled from task accuracy. In particular, within the Qwen3.5 family, GSM8K accuracy scales from $41\%$ ($0.8$B) to $96.8\%$ ($397$B-A$17$B), yet genuine follow-up rates under deterministic generation remain near zero. In contrast, higher temperature sampling reveals interaction awareness is latent with follow up rates reaching $22\%$. Controlled perturbations validate that the proposed probe measures a real property of the model, and collaboration-oriented post-training on Qwen3.5-2B demonstrates an increase in follow-up rates. Our results show that user-turn generation captures a dimension of LLM behavior, interaction awareness, that is unexplored and invisible with current assistant-only benchmarks.

标签

交互意识 LLM评估 用户回复生成

arXiv 分类

cs.AI