LLM Reasoning 相关度: 9/10

Not All Queries Need Deep Thought: CoFiCot for Adaptive Coarse-to-fine Stateful Refinement

Dongxu Zhang, Hongqiang Lin, Yiding Sun, Pengyu Wang, Qirui Wang, Ning Yang, Jihua Zhu
arXiv: 2603.08251v1 发布: 2026-03-09 更新: 2026-03-09

AI 摘要

CoFiCot提出了一种自适应的粗到细推理框架,动态调整LLM的推理策略。

主要贡献

  • 提出CoFiCot框架,动态调整LLM推理资源
  • 使用多指标分类器评估问题难度
  • 引入状态序列传播和过程奖励模型

方法论

通过多指标分类器评估问题难度,对简单问题高效聚合,对复杂问题使用基于历史校正的循环迭代,并使用PRM优化。

原文摘要

Scaling test-time computation enhances LLM reasoning ability but faces a uniform computation paradox. Allocating identical resources leads to over-correction on simple tasks and insufficient refinement on complex ones. To address this, we propose CoFiCot, a coarse-to-fine adaptive framework that dynamically tailors inference strategies to problem difficulty. Specifically, we implement a multi-metric classifier that triages queries by synthesizing semantic entropy, consensus reliability, and predicted reasoning depth . This enables a differentiated refinement stage that applies efficient aggregation for simple queries while routing complex ones to a context-aware correction loop . We formalize correction as a stateful sequential propagation process , where each repair is strictly conditioned on the verified history of prior rectifications. By integrating Process Reward Models (PRMs) within this state-dependent trajectory, CoFiCot effectively bridges the gap between granular error localization and global logical coherence, preventing the context fragmentation typical of stateless refinement methods.

标签

LLM 推理 自适应 Chain-of-Thought

arXiv 分类

cs.CL