Agent Tuning & Optimization 相关度: 7/10

Efficiently Aligning Draft Models via Parameter- and Data-Efficient Adaptation

Luxi Lin, Zhihang Lin, Zhanpeng Zeng, Yuhao Chen, Qingyu Zhang, Jixiang Luo, Xuelong Li, Rongrong Ji
arXiv: 2603.09527v1 发布: 2026-03-10 更新: 2026-03-10

AI 摘要

EDA通过参数高效适配和数据再生策略,提升精调LLM的推测解码性能,降低训练成本。

主要贡献

  • 提出高效草稿模型适配框架EDA
  • 设计解耦架构,分离共享和特定目标组件
  • 引入数据再生策略,提升训练与解码对齐性

方法论

EDA利用解耦架构、数据再生和样本选择,对草稿模型进行参数和数据高效适配。

原文摘要

Speculative decoding accelerates LLM inference but suffers from performance degradation when target models are fine-tuned for specific domains. A naive solution is to retrain draft models for every target model, which is costly and inefficient. To address this, we introduce a parameter- and data-efficient framework named Efficient Draft Adaptation, abbreviated as EDA, for efficiently adapting draft models. EDA introduces three innovations: (1) a decoupled architecture that utilizes shared and private components to model the shared and target-specific output distributions separately, enabling parameter-efficient adaptation by updating only the lightweight private component;(2) a data regeneration strategy that utilizes the fine-tuned target model to regenerate training data, thereby improving the alignment between training and speculative decoding, leading to higher average acceptance length;(3) a sample selection mechanism that prioritizes high-value data for efficient adaptation. Our experiments show that EDA effectively restores speculative performance on fine-tuned models, achieving superior average acceptance lengths with significantly reduced training costs compared to full retraining. Code is available at https://github.com/Lyn-Lucy/Efficient-Draft-Adaptation.

标签

LLM Speculative Decoding Fine-tuning Parameter-efficient Learning

arXiv 分类

cs.LG cs.AI