Agent Tuning & Optimization 相关度: 6/10

ANCRe: Adaptive Neural Connection Reassignment for Efficient Depth Scaling

Yilang Zhang, Bingcong Li, Niao He, Georgios B. Giannakis
arXiv: 2602.09009v1 发布: 2026-02-09 更新: 2026-02-09

AI 摘要

提出自适应神经连接重分配(ANCRe)框架,优化残差连接,提升深度网络的效率。

主要贡献

  • 提出ANCRe框架,自适应学习残差连接
  • 证明残差连接布局影响收敛速度
  • 在多种模型上验证了ANCRe的有效性

方法论

通过参数化和学习残差连接性,ANCRe框架自适应地重新分配残差连接,开销极小。

原文摘要

Scaling network depth has been a central driver behind the success of modern foundation models, yet recent investigations suggest that deep layers are often underutilized. This paper revisits the default mechanism for deepening neural networks, namely residual connections, from an optimization perspective. Rigorous analysis proves that the layout of residual connections can fundamentally shape convergence behavior, and even induces an exponential gap in convergence rates. Prompted by this insight, we introduce adaptive neural connection reassignment (ANCRe), a principled and lightweight framework that parameterizes and learns residual connectivities from the data. ANCRe adaptively reassigns residual connections with negligible computational and memory overhead ($<1\%$), while enabling more effective utilization of network depth. Extensive numerical tests across pre-training of large language models, diffusion models, and deep ResNets demonstrate consistently accelerated convergence, boosted performance, and enhanced depth efficiency over conventional residual connections.

标签

深度学习 残差连接 优化 神经网络 深度缩放

arXiv 分类

cs.LG cs.AI