LLM Reasoning 相关度: 8/10

Universal YOCO for Efficient Depth Scaling

Yutao Sun, Li Dong, Tianzhu Ye, Shaohan Huang, Jianyong Wang, Furu Wei
arXiv: 2604.01220v1 发布: 2026-04-01 更新: 2026-04-01

AI 摘要

YOCO-U结合YOCO和递归计算,提升LLM的推理深度和效率,同时保持低开销和全局KV缓存。

主要贡献

  • 提出 Universal YOCO (YOCO-U) 架构
  • 结合YOCO和递归计算,提升推理效率
  • 实现了常数全局KV缓存和线性预填充

方法论

设计了Universal Self-Decoder,通过参数共享进行多次迭代,并将迭代限制在浅层高效注意力层。

原文摘要

The rise of test-time scaling has remarkably boosted the reasoning and agentic proficiency of Large Language Models (LLMs). Yet, standard Transformers struggle to scale inference-time compute efficiently, as conventional looping strategies suffer from high computational overhead and a KV cache that inflates alongside model depth. We present Universal YOCO (YOCO-U), which combines the YOCO decoder-decoder architecture with recursive computation to achieve a synergistic effect greater than either alone. Built on the YOCO framework, YOCO-U implements a Universal Self-Decoder that performs multiple iterations via parameter sharing, while confining the iterative process to shallow, efficient-attention layers. This combination yields a favorable capability-efficiency tradeoff that neither YOCO nor recursion achieves independently. The YOCO architecture provides a constant global KV cache and linear pre-filling, while partial recursion enhances representational depth with limited overhead. Together, YOCO-U improves token utility and scaling behavior while maintaining efficient inference. Empirical results confirm that YOCO-U remains highly competitive in general and long-context benchmarks, demonstrating that the integration of efficient-attention architectures and recursive computation is a promising direction for scalable LLMs.

标签

LLM 深度缩放 推理效率 递归计算 注意力机制

arXiv 分类

cs.CL