LLM Reasoning 相关度: 9/10

Quantifying the Necessity of Chain of Thought through Opaque Serial Depth

Jonah Brown-Cohen, David Lindner, Rohin Shah
arXiv: 2603.09786v1 发布: 2026-03-10 更新: 2026-03-10

AI 摘要

论文通过“不透明串行深度”量化了LLM进行外部化推理(如CoT)的必要性。

主要贡献

  • 提出了“不透明串行深度”的概念
  • 计算了Gemma 3模型的不透明串行深度上限
  • 开源了计算神经网络不透明串行深度的自动化方法

方法论

形式化不透明串行深度,并通过计算和实验证明其与模型结构的关系,开源相关工具。

原文摘要

Large language models (LLMs) tend to externalize their reasoning in their chain of thought, making the chain of thought a good target for monitoring. This is partially an inherent feature of the Transformer architecture: sufficiently long serial cognition must pass through the chain of thought (Korbak et al., 2025). We formalize this argument through the notion of opaque serial depth, given by the length of the longest computation that can be done without the use of interpretable intermediate steps like chain of thought. Given this formalization, we compute numeric upper bounds on the opaque serial depth of Gemma 3 models, as well as asymptotic results for additional architectures beyond standard LLMs. We also open-source an automated method that can calculate upper bounds on the opaque serial depth of arbitrary neural networks, and use it to demonstrate that Mixture-of-Experts models likely have lower depth than dense models. Overall, our results suggest that opaque serial depth is a useful tool for understanding the potential for models to do significant reasoning that is not externalized.

标签

LLM Reasoning Chain-of-Thought Interpretability Transformer

arXiv 分类

cs.AI