Agent Tuning & Optimization 相关度: 7/10

OmniMoE: An Efficient MoE by Orchestrating Atomic Experts at Scale

Jingze Shi, Zhangyang Peng, Yizhang Zhu, Yifan Wu, Guang Liu, Yuyu Luo
arXiv: 2602.05711v1 发布: 2026-02-05 更新: 2026-02-05

AI 摘要

OmniMoE通过原子专家和系统算法协同设计,实现了高效细粒度MoE,显著提升了推理速度和准确性。

主要贡献

  • 提出向量级原子专家概念
  • 设计笛卡尔积路由,降低路由复杂度
  • 提出专家中心调度,优化内存访问

方法论

通过原子专家最大化模型容量,采用系统算法协同设计,优化路由和内存访问,实现高效MoE。

原文摘要

Mixture-of-Experts (MoE) architectures are evolving towards finer granularity to improve parameter efficiency. However, existing MoE designs face an inherent trade-off between the granularity of expert specialization and hardware execution efficiency. We propose OmniMoE, a system-algorithm co-designed framework that pushes expert granularity to its logical extreme. OmniMoE introduces vector-level Atomic Experts, enabling scalable routing and execution within a single MoE layer, while retaining a shared dense MLP branch for general-purpose processing. Although this atomic design maximizes capacity, it poses severe challenges for routing complexity and memory access. To address these, OmniMoE adopts a system-algorithm co-design: (i) a Cartesian Product Router that decomposes the massive index space to reduce routing complexity from O(N) to O(sqrt(N)); and (ii) Expert-Centric Scheduling that inverts the execution order to turn scattered, memory-bound lookups into efficient dense matrix operations. Validated on seven benchmarks, OmniMoE (with 1.7B active parameters) achieves 50.9% zero-shot accuracy across seven benchmarks, outperforming coarse-grained (e.g., DeepSeekMoE) and fine-grained (e.g., PEER) baselines. Crucially, OmniMoE reduces inference latency from 73ms to 6.7ms (a 10.9-fold speedup) compared to PEER, demonstrating that massive-scale fine-grained MoE can be fast and accurate. Our code is open-sourced at https://github.com/flash-algo/omni-moe.

标签

MoE 专家混合模型 参数效率 推理加速

arXiv 分类

cs.CL cs.AI