LLM Reasoning 相关度: 6/10

On the Role of Depth in the Expressivity of RNNs

Maude Lizaire, Michael Rizvi-Martel, Éric Dupuis, Guillaume Rabusseau
arXiv: 2604.02201v1 发布: 2026-04-02 更新: 2026-04-02

AI 摘要

该论文从理论和实验上证明了深度能够有效提升RNN的记忆容量和表达能力。

主要贡献

  • 证明深度能有效提升RNN的记忆容量
  • 揭示深度如何增强RNN的表达能力
  • 分析了2RNNs的特性并表明乘性交互不可替代

方法论

通过数学证明和实验验证,分析了深度和乘性交互对RNN表达能力的影响。

原文摘要

The benefits of depth in feedforward neural networks are well known: composing multiple layers of linear transformations with nonlinear activations enables complex computations. While similar effects are expected in recurrent neural networks (RNNs), it remains unclear how depth interacts with recurrence to shape expressive power. Here, we formally show that depth increases RNNs' memory capacity efficiently with respect to the number of parameters, thus enhancing expressivity both by enabling more complex input transformations and improving the retention of past information. We broaden our analysis to 2RNNs, a generalization of RNNs with multiplicative interactions between inputs and hidden states. Unlike RNNs, which remain linear without nonlinear activations, 2RNNs perform polynomial transformations whose maximal degree grows with depth. We further show that multiplicative interactions cannot, in general, be replaced by layerwise nonlinearities. Finally, we validate these insights empirically on synthetic and real-world tasks.

标签

RNN 深度学习 表达能力 记忆容量 递归神经网络

arXiv 分类

cs.LG