Command Palette
Search for a command to run...
Rishabh Dabral; Anurag Mundhada; Uday Kusupati; Safeer Afaque; Abhishek Sharma; Arjun Jain

Abstract
3D human pose estimation from a single image is a challenging problem, especially for in-the-wild settings due to the lack of 3D annotated data. We propose two anatomically inspired loss functions and use them with a weakly-supervised learning framework to jointly learn from large-scale in-the-wild 2D and indoor/synthetic 3D data. We also present a simple temporal network that exploits temporal and structural cues present in predicted pose sequences to temporally harmonize the pose estimations. We carefully analyze the proposed contributions through loss surface visualizations and sensitivity analysis to facilitate deeper understanding of their working mechanism. Our complete pipeline improves the state-of-the-art by 11.8% and 12% on Human3.6M and MPI-INF-3DHP, respectively, and runs at 30 FPS on a commodity graphics card.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| 3d-human-pose-estimation-on-3dpw | TP-Net | PA-MPJPE: 92.2 |
| 3d-human-pose-estimation-on-human36m | TP-Net | Average MPJPE (mm): 52.1 PA-MPJPE: 36.3 |
| monocular-3d-human-pose-estimation-on-human3 | TP-Net | Average MPJPE (mm): 52.1 Frames Needed: 20 Need Ground Truth 2D Pose: No Use Video Sequence: Yes |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.