Agent Tuning & Optimization 相关度: 6/10

MONET: Modeling and Optimization of neural NEtwork Training from Edge to Data Centers

Jérémy Morlier, Robin Geens, Stef Cuyckens, Arne Symons, Marian Verhelst, Vincent Gripon, Mathieu Léonardon
arXiv: 2603.15002v1 发布: 2026-03-16 更新: 2026-03-16

AI 摘要

MONET建模神经网络训练过程,优化异构数据流加速器上的训练效率。

主要贡献

  • 提出MONET框架,用于建模异构数据流加速器上的神经网络训练
  • 利用MONET探索ResNet-18和GPT-2的硬件架构设计空间
  • 使用MONET优化层融合配置和激活检查点

方法论

MONET基于Stream框架,通过实验验证,建模异构加速器上神经网络的训练过程,并利用遗传算法进行优化。

原文摘要

While hardware-software co-design has significantly improved the efficiency of neural network inference, modeling the training phase remains a critical yet underexplored challenge. Training workloads impose distinct constraints, particularly regarding memory footprint and backpropagation complexity, which existing inference-focused tools fail to capture. This paper introduces MONET, a framework designed to model the training of neural networks on heterogeneous dataflow accelerators. MONET builds upon Stream, an experimentally verified framework that that models the inference of neural networks on heterogeneous dataflow accelerators with layer fusion. Using MONET, we explore the design space of ResNet-18 and a small GPT-2, demonstrating the framework's capability to model training workflows and find better hardware architectures. We then further examine problems that become more complex in neural network training due to the larger design space, such as determining the best layer-fusion configuration. Additionally, we use our framework to find interesting trade-offs in activation checkpointing, with the help of a genetic algorithm. Our findings highlight the importance of a holistic approach to hardware-software co-design for scalable and efficient deep learning deployment.

标签

硬件加速 神经网络训练 模型优化 硬件软件协同设计

arXiv 分类

cs.LG