3260 papers • 126 benchmarks • 313 datasets
This task aims to solve root-relative 3D multi-person pose estimation (person-centric coordinate system). No ground truth human bounding box and human root joint coordinates are used during testing stage. ( Image credit: RootNet )
(Image credit: Papersgraph)
These leaderboards are used to track progress in 3d-multi-person-pose-estimation-root-relative-1
Use these libraries to find 3d-multi-person-pose-estimation-root-relative-1 models and implementations
No subtasks available.
A weakly-supervised transfer learning method that uses mixed 2D and 3D labels in a unified deep neutral network that presents two-stage cascaded structure to regularize the 3D pose prediction, which is effective in the absence of ground truth depth labels.
This work proposes a new single-shot method for multi-person 3D pose estimation in general scenes from a monocular RGB camera which uses novel occlusion-robust pose-maps (ORPM) which enable full body pose inference even under strong partial occlusions by other people and objects in the scene.
This work firstly proposes a fully learning-based, camera distance-aware top-down approach for 3D multi-person pose estimation from a single RGB image, which achieves comparable results with the state-of-the-art 3D single- person pose estimation models without any groundtruth information and significantly outperforms previous 3DMulti-Person pose estimation methods on publicly available datasets.
This work introduces a network that can be trained with additional RGB-D images in a weakly supervised fashion, and achieves state-of-the-art results on the MuPoTS-3D dataset by a considerable margin.
The Human Depth Estimation Network (HDNet), an end-to-end framework for absolute root joint localization in the camera coordinate space, is proposed and shown to outperform the previous state-of-the-art consistently under multiple evaluation metrics.
A novel system that first regresses a set of 2.5D representations of body parts and then reconstructs the 3D absolute poses based on these 2.
This work presents an energy minimization approach to generate smooth, valid trajectories in time, bridging gaps in visibility in multi-person pose estimation, and shows that it is better than other interpolation based approaches and achieves state of the art results.
This work proposes a novel framework integrating graph convolutional networks (GCNs), which unlike the existing GCN, is based on a directed graph that employs the 2D pose estimator's confidence scores to improve the pose estimation results.
This work proposes a two-person pose discriminator that enforces natural two- person interactions and applies a semi-supervised method to overcome the 3D ground-truth data scarcity.
The proposed Distribution-Aware Single-stage DAS model simultaneously localizes person positions and their corresponding body joints in the 3D camera space in a one-pass manner, leading to a simplified pipeline with enhanced efficiency and its stat-of-the-art accuracy for multi-person 3D pose estimation.
Adding a benchmark result helps the community track progress.