LLM Reasoning 相关度: 9/10

NoRA: Breaking the Linear Ceiling of Low-Rank Adaptation via Manifold Expansion

Hung-Hsuan Chen
arXiv: 2602.22911v1 发布: 2026-02-26 更新: 2026-02-26

AI 摘要

NoRA通过引入非线性机制,突破LoRA的线性瓶颈,提升参数效率,在复杂推理任务中表现更优。

主要贡献

  • 提出NoRA,一种非线性低秩适配方法
  • 通过SiLU门控和结构化Dropout实现流形扩展
  • 实验证明NoRA在推理任务上优于LoRA,具有更高的频谱效率

方法论

NoRA使用权重级并行适配器,通过SiLU门控和结构化Dropout注入非线性,扩展流形空间,并使用SVD进行机制分析。

原文摘要

Low-Rank Adaptation (LoRA) dominates parameter-efficient fine-tuning (PEFT). However, it faces a critical ``linear ceiling'' in complex reasoning tasks: simply increasing the rank yields diminishing returns due to intrinsic linear constraints. We introduce NoRA (Non-linear Rank Adaptation), a weight-level parallel adapter that injects SiLU gating and structural dropout to induce manifold expansion. On the SlimOrca benchmark, NoRA breaks this linear barrier: NoRA remarkably at rank 64 (PPL 3.89) outperforms LoRA at rank 512 (PPL 3.90), demonstrating superior spectral efficiency. This advantage generalizes to mathematical reasoning, where NoRA achieves a perplexity of 1.97 on MathInstruct, significantly surpassing LoRA's saturation point of 2.07. Mechanism analysis via Singular Value Decomposition (SVD) confirms that NoRA activates the dormant tail of the singular value spectrum, effectively preventing the rank collapse observed in linear methods.

标签

Low-Rank Adaptation Parameter-Efficient Fine-Tuning Non-linear Adaptation Manifold Expansion

arXiv 分类

cs.LG cs.AI cs.CL