菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-26
📄 Abstract - Latent Gaussian Splatting for 4D Panoptic Occupancy Tracking

Capturing 4D spatiotemporal surroundings is crucial for the safe and reliable operation of robots in dynamic environments. However, most existing methods address only one side of the problem: they either provide coarse geometric tracking via bounding boxes, or detailed 3D structures like voxel-based occupancy that lack explicit temporal association. In this work, we present Latent Gaussian Splatting for 4D Panoptic Occupancy Tracking (LaGS) that advances spatiotemporal scene understanding in a holistic direction. Our approach incorporates camera-based end-to-end tracking with mask-based multi-view panoptic occupancy prediction, and addresses the key challenge of efficiently aggregating multi-view information into 3D voxel grids via a novel latent Gaussian splatting approach. Specifically, we first fuse observations into 3D Gaussians that serve as a sparse point-centric latent representation of the 3D scene, and then splat the aggregated features onto a 3D voxel grid that is decoded by a mask-based segmentation head. We evaluate LaGS on the Occ3D nuScenes and Waymo datasets, achieving state-of-the-art performance for 4D panoptic occupancy tracking. We make our code available at this https URL.

顶级标签: computer vision robotics systems
详细标签: 4d scene understanding panoptic occupancy tracking gaussian splatting dynamic environments multi-view fusion 或 搜索:

用于4D全景占用跟踪的潜在高斯泼溅方法 / Latent Gaussian Splatting for 4D Panoptic Occupancy Tracking


1️⃣ 一句话总结

这项研究提出了一种名为LaGS的新方法,它通过结合相机跟踪和全景占用预测,并利用创新的‘潜在高斯泼溅’技术高效整合多视角信息,实现了对动态环境中物体(如车辆、行人)的精确4D(三维空间加时间)追踪和语义分割,性能在主流数据集上达到领先水平。

源自 arXiv: 2602.23172