Fluid Representations in Reasoning Models
AI 摘要
研究表明,推理模型通过上下文token表示的动态调整实现抽象结构信息的有效处理和问题解决。
主要贡献
- 发现推理模型在推理过程中改进内部的动作和概念表示
- 证明了模型会发展出专注于结构的抽象编码
- 建立了这些适应性能提升问题解决的因果关系
方法论
通过对QwQ-32B模型在Mystery Blocksworld上的分析,结合引导实验,揭示token表示在推理过程中的演化。
原文摘要
Reasoning language models, which generate long chains of thought, dramatically outperform non-reasoning language models on abstract problems. However, the internal model mechanisms that allow this superior performance remain poorly understood. We present a mechanistic analysis of how QwQ-32B - a model specifically trained to produce extensive reasoning traces - process abstract structural information. On Mystery Blocksworld - a semantically obfuscated planning domain - we find that QwQ-32B gradually improves its internal representation of actions and concepts during reasoning. The model develops abstract encodings that focus on structure rather than specific action names. Through steering experiments, we establish causal evidence that these adaptations improve problem solving: injecting refined representations from successful traces boosts accuracy, while symbolic representations can replace many obfuscated encodings with minimal performance loss. We find that one of the factors driving reasoning model performance is in-context refinement of token representations, which we dub Fluid Reasoning Representations.