Diffusion LLMs can think EoS-by-EoS
AI 摘要
扩散LLM通过填充EoS token进行隐藏计算,从而提升复杂推理能力。
主要贡献
- 发现扩散LLM利用EoS token进行推理
- 验证了EoS token在扩散LLM中的隐藏计算作用
- 提出了EoS-by-EoS的扩散LLM推理机制
方法论
通过可控提示实验和因果干预实验,分析EoS token对LLM推理能力的影响,并验证其隐藏计算作用。
原文摘要
Diffusion LLMs have been proposed as an alternative to autoregressive LLMs, excelling especially at complex reasoning tasks with interdependent sub-goals. Curiously, this is particularly true if the generation length, i.e., the number of tokens the model has to output, is set to a much higher value than is required for providing the correct answer to the task, and the model pads its answer with end-of-sequence (EoS) tokens. We hypothesize that diffusion models think EoS-by-EoS, that is, they use the representations of EoS tokens as a hidden scratchpad, which allows them to solve harder reasoning problems. We experiment with the diffusion models LLaDA1.5, LLaDA2.0-mini, and Dream-v0 on the tasks Addition, Entity Tracking, and Sudoku. In a controlled prompting experiment, we confirm that adding EoS tokens improves the LLMs' reasoning capabilities. To further verify whether they serve as space for hidden computations, we patch the hidden states of the EoS tokens with those of a counterfactual generation, which frequently changes the generated output to the counterfactual. The success of the causal intervention underscores that the EoS tokens, which one may expect to be devoid of meaning, carry information on the problem to solve. The behavioral experiments and the causal interventions indicate that diffusion LLMs can indeed think EoS-by-EoS.