菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-29
📄 Abstract - How to Guide Your Flow: Few-Step Alignment via Flow Map Reward Guidance

In generative modeling, we often wish to produce samples that maximize a user-specified reward such as aesthetic quality or alignment with human preferences, a problem known as guidance. Despite their widespread use, existing guidance methods either require expensive multi-particle, many-step schemes or rely on poorly understood approximations. We reformulate guidance as a deterministic optimal control problem, yielding a hierarchy of algorithms that subsumes existing approaches at the coarsest level. We show that the flow map, an object of significant recent interest for its role in fast inference, arises naturally in the optimal solution. Based on this observation, we propose Flow Map Reward Guidance (FMRG): a training-free, single-trajectory framework that uses the flow map to both integrate and guide the flow. At text-to-image scale, FMRG matches or surpasses baselines across inverse problems, style transfer, human preferences, and VLM rewards with as few as 3 NFEs, giving at least an order-of-magnitude speedup in comparison to prior state of the art.

顶级标签: machine learning aigc model training
详细标签: flow map reward guidance few-step inference text-to-image optimal control 或 搜索:

如何引导你的生成流:基于流图奖励引导的极简步数对齐方法 / How to Guide Your Flow: Few-Step Alignment via Flow Map Reward Guidance


1️⃣ 一句话总结

本文提出了一种名为流图奖励引导(FMRG)的新方法,通过将生成过程重构成最优控制问题,利用流图实现无需额外训练、仅需单条轨迹且极少量计算步数(低至3步)的快速高质量生成,显著提升了文本到图像任务中的风格迁移、逆问题求解和对齐人类偏好的效率。

源自 arXiv: 2604.27147