Fast and Scalable Analytical Diffusion
AI 摘要
提出了一种高效的Analytical Diffusion模型GoldDiff,通过动态选择“Golden Subset”加速推理,显著提升了模型的可扩展性。
主要贡献
- 发现后验渐进集中现象
- 提出Dynamic Time-Aware Golden Subset Diffusion (GoldDiff)框架
- 实现了Analytical Diffusion模型在ImageNet-1K上的成功扩展
方法论
利用信号噪声比提高时,有效支持集从全局流形收缩到局部邻域的特性,动态选择关键数据子集进行推理。
原文摘要
Analytical diffusion models offer a mathematically transparent path to generative modeling by formulating the denoising score as an empirical-Bayes posterior mean. However, this interpretability comes at a prohibitive cost: the standard formulation necessitates a full-dataset scan at every timestep, scaling linearly with dataset size. In this work, we present the first systematic study addressing this scalability bottleneck. We challenge the prevailing assumption that the entire training data is necessary, uncovering the phenomenon of Posterior Progressive Concentration: the effective golden support of the denoising score is not static but shrinks asymptotically from the global manifold to a local neighborhood as the signal-to-noise ratio increases. Capitalizing on this, we propose Dynamic Time-Aware Golden Subset Diffusion (GoldDiff), a training-free framework that decouples inference complexity from dataset size. Instead of static retrieval, GoldDiff uses a coarse-to-fine mechanism to dynamically pinpoint the ''Golden Subset'' for inference. Theoretically, we derive rigorous bounds guaranteeing that our sparse approximation converges to the exact score. Empirically, GoldDiff achieves a $\bf 71 \times$ speedup on AFHQ while matching or achieving even better performance than full-scan baselines. Most notably, we demonstrate the first successful scaling of analytical diffusion to ImageNet-1K, unlocking a scalable, training-free paradigm for large-scale generative modeling.