菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-17
📄 Abstract - WildDepth: A Multimodal Dataset for 3D Wildlife Perception and Depth Estimation

Depth estimation and 3D reconstruction have been extensively studied as core topics in computer vision. Starting from rigid objects with relatively simple geometric shapes, such as vehicles, the research has expanded to address general objects, including challenging deformable objects, such as humans and animals. However, for the animal, in particular, the majority of existing models are trained based on datasets without metric scale, which can help validate image-only models. To address this limitation, we present WildDepth, a multimodal dataset and benchmark suite for depth estimation, behavior detection, and 3D reconstruction from diverse categories of animals ranging from domestic to wild environments with synchronized RGB and LiDAR. Experimental results show that the use of multi-modal data improves depth reliability by up to 10% RMSE, while RGB-LiDAR fusion enhances 3D reconstruction fidelity by 12% in Chamfer distance. By releasing WildDepth and its benchmarks, we aim to foster robust multimodal perception systems that generalize across domains.

顶级标签: computer vision multi-modal data
详细标签: depth estimation 3d reconstruction dataset lidar wildlife perception 或 搜索:

WildDepth:用于3D野生动物感知与深度估计的多模态数据集 / WildDepth: A Multimodal Dataset for 3D Wildlife Perception and Depth Estimation


1️⃣ 一句话总结

这篇论文提出了一个名为WildDepth的新型多模态数据集,它结合了彩色图像和激光雷达数据,专门用于提升对动物进行三维感知、深度估计和行为检测的准确性,实验表明该数据集能显著提高相关任务的性能。

源自 arXiv: 2603.16816