Agent Tuning & Optimization 相关度: 6/10

Joint Hardware-Workload Co-Optimization for In-Memory Computing Accelerators

Olga Krestinskaya, Mohammed E. Fouda, Ahmed Eltawil, Khaled N. Salama
arXiv: 2603.03880v1 发布: 2026-03-04 更新: 2026-03-04

AI 摘要

提出一种面向多工作负载的片上内存计算加速器联合软硬件协同优化框架,显著提升通用性。

主要贡献

  • 提出了一种基于优化进化算法的联合软硬件协同优化框架。
  • 该框架能够显式地捕获跨工作负载的权衡,优化通用IMC设计。
  • 在RRAM和SRAM两种架构上进行了验证,证明了其鲁棒性和适应性。

方法论

利用优化进化算法,对片上内存计算加速器的硬件架构和工作负载进行联合优化,以减少EDAP。

原文摘要

Software-hardware co-design is essential for optimizing in-memory computing (IMC) hardware accelerators for neural networks. However, most existing optimization frameworks target a single workload, leading to highly specialized hardware designs that do not generalize well across models and applications. In contrast, practical deployment scenarios require a single IMC platform that can efficiently support multiple neural network workloads. This work presents a joint hardware-workload co-optimization framework based on an optimized evolutionary algorithm for designing generalized IMC accelerator architectures. By explicitly capturing cross-workload trade-offs rather than optimizing for a single model, the proposed approach significantly reduces the performance gap between workload-specific and generalized IMC designs. The framework is evaluated on both RRAM- and SRAM-based IMC architectures, demonstrating strong robustness and adaptability across diverse design scenarios. Compared to baseline methods, the optimized designs achieve energy-delay-area product (EDAP) reductions of up to 76.2% and 95.5% when optimizing across a small set (4 workloads) and a large set (9 workloads), respectively. The source code of the framework is available at https://github.com/OlgaKrestinskaya/JointHardwareWorkloadOptimizationIMC.

标签

In-Memory Computing Hardware-Workload Co-design Evolutionary Algorithm

arXiv 分类

cs.AR cs.AI cs.ET cs.NE eess.SY