LLM Reasoning 相关度: 8/10

Brainstacks: Cross-Domain Cognitive Capabilities via Frozen MoE-LoRA Stacks for Continual LLM Learning

Mohammad R. Abu Ayyash
arXiv: 2604.01152v1 发布: 2026-04-01 更新: 2026-04-01

AI 摘要

Brainstacks提出了一种模块化的持续学习架构,通过冻结MoE-LoRA堆栈实现跨领域认知能力。

主要贡献

  • 提出了Brainstacks架构,用于持续多领域微调LLM
  • 利用MoE-LoRA实现更快的收敛速度
  • 发现领域堆栈编码可转移的认知原语,而非领域特定知识

方法论

采用冻结的MoE-LoRA堆栈,通过残差提升、空空间投影和基于结果的元路由实现持续学习和跨领域组合。

原文摘要

We present Brainstacks, a modular architecture for continual multi-domain fine-tuning of large language models that packages domain expertise as frozen adapter stacks composing additively on a shared frozen base at inference. Five interlocking components: (1) MoE-LoRA with Shazeer-style noisy top-2 routing across all seven transformer projections under QLoRA 4-bit quantization with rsLoRA scaling; (2) an inner loop performing residual boosting by freezing trained stacks and adding new ones; (3) an outer loop training sequential domain-specific stacks with curriculum-ordered dependencies; (4) null-space projection via randomized SVD constraining new stacks to subspaces orthogonal to prior directions, achieving zero forgetting in isolation; (5) an outcome-based sigmoid meta-router trained on empirically discovered domain-combination targets that selectively weights stacks, enabling cross-domain composition. Two boundary experiments: (6) PSN pretraining on a randomly initialized model; (7) per-domain RL (DPO/GRPO) validating compatibility with post-SFT alignment. Validated on TinyLlama-1.1B (4 domains, 9 stacks) and Gemma 3 12B IT (5 domains, 10 stacks), MoE-LoRA achieves 2.5x faster convergence than parameter-matched single LoRA, residual boosting breaks through the single-stack ceiling, and the routed system recovers generation quality destroyed by ungated stack accumulation. The central finding: the outcome-based router discovers that domain stacks encode transferable cognitive primitives (instruction-following clarity, numerical reasoning, procedural logic, chain-of-thought structure) rather than domain-specific knowledge, with medical prompts routing to chat+math stacks in 97% of cases despite zero medical data in those stacks.

标签

持续学习 LoRA MoE 跨领域 认知能力

arXiv 分类

cs.CL cs.AI