InterMesh:显式交互感知的端到端多人人体网格恢复 / InterMesh: Explicit Interaction-Aware End-to-End Multi-Person Human Mesh Recovery
1️⃣ 一句话总结
InterMesh 提出了一种新方法,通过在现有框架中加入显式的人-物和人-人交互信息,让计算机更准确地从单张图片中恢复多人身体的3D形状和姿态,并在多个标准测试集上显著提升了精度。
Humans constantly interact with their surroundings. Existing end-to-end multi-person human mesh recovery methods, typically based on the DETR framework, capture inter-human relationships through self-attention across all human queries. However, these approaches model interactions only implicitly and lack explicit reasoning about how humans interact with objects and with each other. In this paper, we propose InterMesh, a simple yet effective framework that explicitly incorporates human-environment interaction information into human mesh recovery pipeline. By leveraging a human-object interaction detector, InterMesh enriches query representations with structured interaction semantics, enabling more accurate pose and shape estimation. We design lightweight modules, Contextual Interaction Encoder and Interaction-Guided Refiner, to integrate these features into existing HMR architectures with minimal overhead. We validate our approach through extensive experiments on 3DPW, MuPoTS, CMU Panoptic, Hi4D, and CHI3D datasets, demonstrating remarkable improvements over state-of-the-art methods. Notably, InterMesh reduces MPJPE by 9.9% on CMU Panoptic and 8.2% on Hi4D, highlighting its effectiveness in scenarios with complex human-object and inter-human interactions.
InterMesh:显式交互感知的端到端多人人体网格恢复 / InterMesh: Explicit Interaction-Aware End-to-End Multi-Person Human Mesh Recovery
InterMesh 提出了一种新方法,通过在现有框架中加入显式的人-物和人-人交互信息,让计算机更准确地从单张图片中恢复多人身体的3D形状和姿态,并在多个标准测试集上显著提升了精度。
源自 arXiv: 2605.04554