菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-16
📄 Abstract - PhysRVG: Physics-Aware Unified Reinforcement Learning for Video Generative Models

Physical principles are fundamental to realistic visual simulation, but remain a significant oversight in transformer-based video generation. This gap highlights a critical limitation in rendering rigid body motion, a core tenet of classical mechanics. While computer graphics and physics-based simulators can easily model such collisions using Newton formulas, modern pretrain-finetune paradigms discard the concept of object rigidity during pixel-level global denoising. Even perfectly correct mathematical constraints are treated as suboptimal solutions (i.e., conditions) during model optimization in post-training, fundamentally limiting the physical realism of generated videos. Motivated by these considerations, we introduce, for the first time, a physics-aware reinforcement learning paradigm for video generation models that enforces physical collision rules directly in high-dimensional spaces, ensuring the physics knowledge is strictly applied rather than treated as conditions. Subsequently, we extend this paradigm to a unified framework, termed Mimicry-Discovery Cycle (MDcycle), which allows substantial fine-tuning while fully preserving the model's ability to leverage physics-grounded feedback. To validate our approach, we construct new benchmark PhysRVGBench and perform extensive qualitative and quantitative experiments to thoroughly assess its effectiveness.

顶级标签: video generation reinforcement learning model training
详细标签: physics-aware generation rigid body motion collision simulation reinforcement learning fine-tuning video benchmark 或 搜索:

PhysRVG:面向视频生成模型的物理感知统一强化学习 / PhysRVG: Physics-Aware Unified Reinforcement Learning for Video Generative Models


1️⃣ 一句话总结

这篇论文首次提出了一种物理感知的强化学习框架,通过直接在视频生成过程中强制执行物理碰撞规则,并引入一个名为‘模仿-发现循环’的统一训练范式,显著提升了生成视频中刚体运动的物理真实感。

源自 arXiv: 2601.11087