LLM Reasoning 相关度: 9/10

Tiny Recursive Reasoning with Mamba-2 Attention Hybrid

Wenlong Wang, Fergal Reid
arXiv: 2602.12078v1 发布: 2026-02-12 更新: 2026-02-12

AI 摘要

该论文探索了将Mamba-2算子融入递归推理模型的可行性,并验证了其在保持推理能力的同时具有性能提升。

主要贡献

  • 验证了Mamba-2算子在递归推理框架中的可行性
  • 发现了Mamba-2混合算子能提升ARC-AGI-1数据集上的性能
  • 为递归算子设计空间提供了新的选择

方法论

用Mamba-2混合算子替换TRM中的Transformer块,保持参数量相似,并在ARC-AGI-1数据集上进行评估。

原文摘要

Recent work on recursive reasoning models like TRM demonstrates that tiny networks (7M parameters) can achieve strong performance on abstract reasoning tasks through latent recursion -- iterative refinement in hidden representation space without emitting intermediate tokens. This raises a natural question about operator choice: Mamba-2's state space recurrence is itself a form of iterative refinement, making it a natural candidate for recursive reasoning -- but does introducing Mamba-2 into the recursive scaffold preserve reasoning capability? We investigate this by replacing the Transformer blocks in TRM with Mamba-2 hybrid operators while maintaining parameter parity (6.83M vs 6.86M parameters). On ARC-AGI-1, we find that the hybrid improves pass@2 (the official metric) by +2.0\% (45.88\% vs 43.88\%) and consistently outperforms at higher K values (+4.75\% at pass@100), whilst maintaining pass@1 parity. This suggests improved candidate coverage -- the model generates correct solutions more reliably -- with similar top-1 selection. Our results validate that Mamba-2 hybrid operators preserve reasoning capability within the recursive scaffold, establishing SSM-based operators as viable candidates in the recursive operator design space and taking a first step towards understanding the best mixing strategies for recursive reasoning.

标签

递归推理 Mamba-2 状态空间模型 抽象推理

arXiv 分类

cs.AI cs.CL