菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-25
📄 Abstract - HGGT: Robust and Flexible 3D Hand Mesh Reconstruction from Uncalibrated Images

Recovering high-fidelity 3D hand geometry from images is a critical task in computer vision, holding significant value for domains such as robotics, animation and VR/AR. Crucially, scalable applications demand both accuracy and deployment flexibility, requiring the ability to leverage massive amounts of unstructured image data from the internet or enable deployment on consumer-grade RGB cameras without complex calibration. However, current methods face a dilemma. While single-view approaches are easy to deploy, they suffer from depth ambiguity and occlusion. Conversely, multi-view systems resolve these uncertainties but typically demand fixed, calibrated setups, limiting their real-world utility. To bridge this gap, we draw inspiration from 3D foundation models that learn explicit geometry directly from visual data. By reformulating hand reconstruction from arbitrary views as a visual-geometry grounded task, we propose a feed-forward architecture that, for the first time in literature, jointly infers 3D hand meshes and camera poses from uncalibrated views. Extensive evaluations show that our approach outperforms state-of-the-art benchmarks and demonstrates strong generalization to uncalibrated, in-the-wild scenarios. Here is the link of our project page: this https URL.

顶级标签: computer vision systems model training
详细标签: 3d reconstruction hand mesh multi-view camera pose estimation uncalibrated images 或 搜索:

HGGT:从非标定图像中实现鲁棒且灵活的三维手部网格重建 / HGGT: Robust and Flexible 3D Hand Mesh Reconstruction from Uncalibrated Images


1️⃣ 一句话总结

这篇论文提出了一种新方法,能够直接从任意角度、未经校准的普通照片中,同时精确地重建出三维手部模型并估算拍摄角度,解决了现有技术要么精度不足、要么部署要求苛刻的难题。

源自 arXiv: 2603.23997