用于加速扩散变换器的关系特征缓存 / Relational Feature Caching for Accelerating Diffusion Transformers
1️⃣ 一句话总结
这篇论文提出了一种名为‘关系特征缓存’的新方法,通过利用神经网络模块输入与输出之间的关系来更准确地预测和缓存中间计算结果,从而在保持生成质量的同时,显著提升了扩散模型(一种图像生成AI)的推理速度。
Feature caching approaches accelerate diffusion transformers (DiTs) by storing the output features of computationally expensive modules at certain timesteps, and exploiting them for subsequent steps to reduce redundant computations. Recent forecasting-based caching approaches employ temporal extrapolation techniques to approximate the output features with cached ones. Although effective, relying exclusively on temporal extrapolation still suffers from significant prediction errors, leading to performance degradation. Through a detailed analysis, we find that 1) these errors stem from the irregular magnitude of changes in the output features, and 2) an input feature of a module is strongly correlated with the corresponding output. Based on this, we propose relational feature caching (RFC), a novel framework that leverages the input-output relationship to enhance the accuracy of the feature prediction. Specifically, we introduce relational feature estimation (RFE) to estimate the magnitude of changes in the output features from the inputs, enabling more accurate feature predictions. We also present relational cache scheduling (RCS), which estimates the prediction errors using the input features and performs full computations only when the errors are expected to be substantial. Extensive experiments across various DiT models demonstrate that RFC consistently outperforms prior approaches significantly. Project page is available at this https URL
用于加速扩散变换器的关系特征缓存 / Relational Feature Caching for Accelerating Diffusion Transformers
这篇论文提出了一种名为‘关系特征缓存’的新方法,通过利用神经网络模块输入与输出之间的关系来更准确地预测和缓存中间计算结果,从而在保持生成质量的同时,显著提升了扩散模型(一种图像生成AI)的推理速度。
源自 arXiv: 2602.19506