菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-18
📄 Abstract - StageVAR: Stage-Aware Acceleration for Visual Autoregressive Models

Visual Autoregressive (VAR) modeling departs from the next-token prediction paradigm of traditional Autoregressive (AR) models through next-scale prediction, enabling high-quality image generation. However, the VAR paradigm suffers from sharply increased computational complexity and running time at large-scale steps. Although existing acceleration methods reduce runtime for large-scale steps, but rely on manual step selection and overlook the varying importance of different stages in the generation process. To address this challenge, we present StageVAR, a systematic study and stage-aware acceleration framework for VAR models. Our analysis shows that early steps are critical for preserving semantic and structural consistency and should remain intact, while later steps mainly refine details and can be pruned or approximated for acceleration. Building on these insights, StageVAR introduces a plug-and-play acceleration strategy that exploits semantic irrelevance and low-rank properties in late-stage computations, without requiring additional training. Our proposed StageVAR achieves up to 3.4x speedup with only a 0.01 drop on GenEval and a 0.26 decrease on DPG, consistently outperforming existing acceleration baselines. These results highlight stage-aware design as a powerful principle for efficient visual autoregressive image generation.

顶级标签: model training computer vision aigc
详细标签: autoregressive models image generation computational efficiency stage-aware acceleration inference optimization 或 搜索:

StageVAR:面向视觉自回归模型的阶段感知加速框架 / StageVAR: Stage-Aware Acceleration for Visual Autoregressive Models


1️⃣ 一句话总结

这篇论文提出了一个名为StageVAR的阶段感知加速框架,它通过分析图像生成过程中不同阶段的重要性,在保持早期关键步骤完整的同时,对后期细节优化步骤进行剪枝或近似计算,从而在不需额外训练的情况下,显著提升了视觉自回归模型的生成速度,且几乎不影响图像质量。

源自 arXiv: 2512.16483