菜单

🤖 系统
📄 Abstract - VIDEOP2R: Video Understanding from Perception to Reasoning

Reinforcement fine-tuning (RFT), a two-stage framework consisting of supervised fine-tuning (SFT) and reinforcement learning (RL) has shown promising results on improving reasoning ability of large language models (LLMs). Yet extending RFT to large video language models (LVLMs) remains challenging. We propose VideoP2R, a novel process-aware video RFT framework that enhances video reasoning by modeling perception and reasoning as distinct processes. In the SFT stage, we develop a three-step pipeline to generate VideoP2R-CoT-162K, a high-quality, process-aware chain-of-thought (CoT) dataset for perception and reasoning. In the RL stage, we introduce a novel process-aware group relative policy optimization (PA-GRPO) algorithm that supplies separate rewards for perception and reasoning. Extensive experiments show that VideoP2R achieves state-of-the-art (SotA) performance on six out of seven video reasoning and understanding benchmarks. Ablation studies further confirm the effectiveness of our process-aware modeling and PA-GRPO and demonstrate that model's perception output is information-sufficient for downstream reasoning.

顶级标签: video model training multi-modal
详细标签: video reasoning reinforcement fine-tuning chain-of-thought process-aware modeling video language models 或 搜索:

📄 论文总结

VIDEOP2R:从感知到推理的视频理解 / VIDEOP2R: Video Understanding from Perception to Reasoning


1️⃣ 一句话总结

本文提出VideoP2R框架,通过将视频理解分为感知和推理两个独立过程进行建模与优化,在多个视频推理基准测试中取得了领先性能。


📄 打开原文 PDF