菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-03
📄 Abstract - ReCamDriving: LiDAR-Free Camera-Controlled Novel Trajectory Video Generation

We propose ReCamDriving, a purely vision-based, camera-controlled novel-trajectory video generation framework. While repair-based methods fail to restore complex artifacts and LiDAR-based approaches rely on sparse and incomplete cues, ReCamDriving leverages dense and scene-complete 3DGS renderings for explicit geometric guidance, achieving precise camera-controllable generation. To mitigate overfitting to restoration behaviors when conditioned on 3DGS renderings, ReCamDriving adopts a two-stage training paradigm: the first stage uses camera poses for coarse control, while the second stage incorporates 3DGS renderings for fine-grained viewpoint and geometric guidance. Furthermore, we present a 3DGS-based cross-trajectory data curation strategy to eliminate the train-test gap in camera transformation patterns, enabling scalable multi-trajectory supervision from monocular videos. Based on this strategy, we construct the ParaDrive dataset, containing over 110K parallel-trajectory video pairs. Extensive experiments demonstrate that ReCamDriving achieves state-of-the-art camera controllability and structural consistency.

顶级标签: computer vision video generation multi-modal
详细标签: novel view synthesis 3d gaussian splatting camera control video generation geometric guidance 或 搜索:

ReCamDriving:一种无需激光雷达、由相机控制的新轨迹视频生成方法 / ReCamDriving: LiDAR-Free Camera-Controlled Novel Trajectory Video Generation


1️⃣ 一句话总结

这篇论文提出了一种仅使用视觉信息(无需激光雷达)就能根据新相机轨迹生成逼真驾驶视频的新方法,它通过一种创新的两阶段训练策略和新的数据集,实现了对相机视角的精确控制和视频结构的高度一致性。


源自 arXiv: 2512.03621