菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-08
📄 Abstract - VDPP: Video Depth Post-Processing for Speed and Scalability

Video depth estimation is essential for providing 3D scene structure in applications ranging from autonomous driving to mixed reality. Current end-to-end video depth models have established state-of-the-art performance. Although current end-to-end (E2E) models have achieved state-of-the-art performance, they function as tightly coupled systems that suffer from a significant adaptation lag whenever superior single-image depth estimators are released. To mitigate this issue, post-processing methods such as NVDS offer a modular plug-and-play alternative to incorporate any evolving image depth model without retraining. However, existing post-processing methods still struggle to match the efficiency and practicality of E2E systems due to limited speed, accuracy, and RGB reliance. In this work, we revitalize the role of post-processing by proposing VDPP (Video Depth Post-Processing), a framework that improves the speed and accuracy of post-processing methods for video depth estimation. By shifting the paradigm from computationally expensive scene reconstruction to targeted geometric refinement, VDPP operates purely on geometric refinements in low-resolution space. This design achieves exceptional speed (>43.5 FPS on NVIDIA Jetson Orin Nano) while matching the temporal coherence of E2E systems, with dense residual learning driving geometric representations rather than full reconstructions. Furthermore, our VDPP's RGB-free architecture ensures true scalability, enabling immediate integration with any evolving image depth model. Our results demonstrate that VDPP provides a superior balance of speed, accuracy, and memory efficiency, making it the most practical solution for real-time edge deployment. Our project page is at this https URL

顶级标签: computer vision video systems
详细标签: video depth estimation post-processing real-time edge deployment geometric refinement 或 搜索:

VDPP:面向速度与可扩展性的视频深度后处理 / VDPP: Video Depth Post-Processing for Speed and Scalability


1️⃣ 一句话总结

这篇论文提出了一种名为VDPP的视频深度后处理框架,它通过专注于低分辨率空间的几何优化而非复杂场景重建,实现了高速、高精度且不依赖RGB信息的视频深度估计,能快速兼容任何最新的单图深度模型,非常适合实时边缘设备部署。

源自 arXiv: 2604.06665