GenHMR: A Breakthrough in 3D Human Pose and Shape Estimation – Accepted at AAAI 2025

Categories: AI4Health News

We are pleased to announce that our paper, “GenHMR: Generative Human Mesh Recovery for Accurate 3D Pose and Shape Estimation,” has been accepted for presentation at AAAI 2025. GenHMR introduces a novel generative framework that surpasses existing methods in parametric 3D human pose and shape (HPS) estimation, offering state-of-the-art accuracy and robustness.

Advancing 3D Human Mesh Recovery

Traditional deterministic and probabilistic methods for 3D pose estimation struggle with depth ambiguity, occlusions, and uncertainty in 2D-to-3D mappings. GenHMR redefines the problem by explicitly modeling uncertainties and leveraging generative modeling techniques to achieve high-fidelity 3D reconstructions.

Key Innovations of GenHMR:

  • Generative HMR Reformulation: Moves beyond deterministic pipelines by modeling uncertainties in 2D-to-3D mapping, effectively addressing depth ambiguities and occlusions.
  • Pose Tokenizer (VQ-VAE): Encodes 3D poses into discrete tokens, ensuring compact, kinematically valid representations for high-quality reconstructions.
  • Image-Conditional Masked Transformer: Learns probabilistic pose distributions by predicting masked pose tokens, capturing complex joint dependencies for enhanced 3D accuracy.
  • Uncertainty-Guided Sampling: Iteratively refines pose predictions, progressively improving accuracy by re-predicting low-confidence tokens during inference.
  • 2D Pose-Guided Refinement: Ensures precise alignment between the reconstructed 3D mesh and 2D pose cues, enhancing final outputs.

Setting a New Standard for 3D Human Mesh Recovery

By combining generative modeling, iterative refinement, and uncertainty quantification, GenHMR outperforms existing methods in both accuracy and robustness, establishing a new benchmark for 3D human pose estimation.

Acknowledgments

This research was made possible through the contributions of Pu Wang, Ekkasit Pinyoanuntapong, Hongfei Xue, Srijan Das, and Chen Chen.

📄 Read the Full Paper: Link
🔗 Project Page: Link

Stay connected for more updates on advancements in human-centric AI and 3D vision research.