Attention-Driven Training-Free Efficiency Enhancement of Diffusion Models
CVPR 2024(2024)
摘要
Diffusion Models (DMs) have exhibited superior performance in generatinghigh-quality and diverse images. However, this exceptional performance comes atthe cost of expensive architectural design, particularly due to the attentionmodule heavily used in leading models. Existing works mainly adopt a retrainingprocess to enhance DM efficiency. This is computationally expensive and notvery scalable. To this end, we introduce the Attention-driven Training-freeEfficient Diffusion Model (AT-EDM) framework that leverages attention maps toperform run-time pruning of redundant tokens, without the need for anyretraining. Specifically, for single-denoising-step pruning, we develop a novelranking algorithm, Generalized Weighted Page Rank (G-WPR), to identifyredundant tokens, and a similarity-based recovery method to restore tokens forthe convolution operation. In addition, we propose a Denoising-Steps-AwarePruning (DSAP) approach to adjust the pruning budget across different denoisingtimesteps for better generation quality. Extensive evaluations show that AT-EDMperforms favorably against prior art in terms of efficiency (e.g., 38.8saving and up to 1.53x speed-up over Stable Diffusion XL) while maintainingnearly the same FID and CLIP scores as the full model. Project webpage:https://atedm.github.io.
更多查看译文
关键词
diffusion model,training-free,efficiency,token pruning,attention map
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
数据免责声明
页面数据均来自互联网公开来源、合作出版商和通过AI技术自动分析结果,我们不对页面数据的有效性、准确性、正确性、可靠性、完整性和及时性做出任何承诺和保证。若有疑问,可以通过电子邮件方式联系我们:report@aminer.cn