3260 papers • 126 benchmarks • 313 datasets
Encoding and reconstruction of 3D point clouds.
(Image credit: Papersgraph)
These leaderboards are used to track progress in 3d-point-cloud-reconstruction-7
No benchmarks available.
Use these libraries to find 3d-point-cloud-reconstruction-7 models and implementations
No subtasks available.
The main contribution is in extending the loss function of YOLO v2 to include the yaw angle, the 3D box center in Cartesian coordinates and the height of the box as a direct regression problem, which enables real-time performance, which is essential for automated driving.
This work introduces a novel differentiable relaxation for point cloud sampling that approximates sampled points as a mixture of points in the primary input cloud and outperforms existing non-learned and learned sampling alternatives.
3D-LMNet, a latent embedding matching approach for 3D reconstruction, is proposed, which outperform state-of-the-art approaches on the task of single-view3D reconstruction on both real and synthetic datasets while generating multiple plausible reconstructions, demonstrating the generalizability and utility of the approach.
It is demonstrated that jointly training for both reconstruction and segmentation leads to improved performance in both the tasks, when compared to training for each task individually.
A novel differentiable projection module, called ‘CAPNet’, is introduced to obtain 2D masks from a predicted 3D point cloud reconstruction, and significantly outperform the existing projection based approaches on a large-scale synthetic dataset.
This work introduces DensePCR, a deep pyramidal network for point cloud reconstruction that hierarchically predicts point clouds of increasing resolution, and proposes an architecture that first predicts a low-resolution point cloud, and then hierarchically increases the resolution by aggregating local and global point features to deform a grid.
A novel dense surfel mapping system that scales well in different environments with only CPU computation using a sparse SLAM system to estimate camera poses, which can fuse intensity images and depth images into a globally consistent model.
This paper proposes an effective and efficient pyramid multi-view stereo (MVS) net with self-adaptive view aggregation for accurate and complete dense point cloud reconstruction and establishes a new state-of-the-art on the DTU dataset with significant improvements in the completeness and overall quality.
This work proposes a physically-based network to recover 3D shape of transparent objects using a few images acquired with a mobile phone camera, under a known but arbitrary environment map and renders a synthetic dataset to encourage the model to learn refractive light transport across different views.
Adding a benchmark result helps the community track progress.