菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-30
📄 Abstract - PhyCo: Learning Controllable Physical Priors for Generative Motion

Modern video diffusion models excel at appearance synthesis but still struggle with physical consistency: objects drift, collisions lack realistic rebound, and material responses seldom match their underlying properties. We present PhyCo, a framework that introduces continuous, interpretable, and physically grounded control into video generation. Our approach integrates three key components: (i) a large-scale dataset of over 100K photorealistic simulation videos where friction, restitution, deformation, and force are systematically varied across diverse scenarios; (ii) physics-supervised fine-tuning of a pretrained diffusion model using a ControlNet conditioned on pixel-aligned physical property maps; and (iii) VLM-guided reward optimization, where a fine-tuned vision-language model evaluates generated videos with targeted physics queries and provides differentiable feedback. This combination enables a generative model to produce physically consistent and controllable outputs through variations in physical attributes-without any simulator or geometry reconstruction at inference. On the Physics-IQ benchmark, PhyCo significantly improves physical realism over strong baselines, and human studies confirm clearer and more faithful control over physical attributes. Our results demonstrate a scalable path toward physically consistent, controllable generative video models that generalize beyond synthetic training environments.

顶级标签: video generation aigc computer vision
详细标签: video diffusion physical priors controllable generation physics consistency controlnet 或 搜索:

PhyCo:学习可控物理先验以生成运动 / PhyCo: Learning Controllable Physical Priors for Generative Motion


1️⃣ 一句话总结

本文提出了一种名为PhyCo的框架,通过结合大规模物理仿真数据集、物理监督的扩散模型微调以及视觉语言模型引导的优化,使视频生成模型能够精确控制物体的摩擦、弹性等物理属性,从而生成物理上更真实、更可控的运动视频。

源自 arXiv: 2604.28169