3260 papers • 126 benchmarks • 313 datasets
Image: Choy et al
(Image credit: Papersgraph)
These leaderboards are used to track progress in 3d-object-reconstruction-7
Use these libraries to find 3d-object-reconstruction-7 models and implementations
The 3D-R2N2 reconstruction framework outperforms the state-of-the-art methods for single view reconstruction, and enables the 3D reconstruction of objects in situations when traditional SFM/SLAM methods fail (because of lack of texture and/or wide baseline).
Experimental results on the ShapeNet and Pix3D benchmarks indicate that the proposed Pix2Vox outperforms state-of-the-arts by a large margin, and the proposed method is 24 times faster than 3D-R2N2 in terms of backward inference time.
An end-to-end deep learning architecture that produces a 3D shape in triangular mesh from a single color image by progressively deforming an ellipsoid, leveraging perceptual features extracted from the input image.
This paper addresses the problem of 3D reconstruction from a single image, generating a straight-forward form of output unorthordox, and designs architecture, loss function and learning paradigm that are novel and effective, capable of predicting multiple plausible 3D point clouds from an input image.
This work shows that extracting 3d hand motion for in-hand scanning effectively facilitates the reconstruction of even featureless and highly symmetric objects and presents an approach that fuses the rich additional information of hands into a 3d reconstruction pipeline, significantly contributing to the state-of-the-art of in- hand scanning.
This paper uses 2D convolutional operations to predict the 3D structure from multiple viewpoints and jointly apply geometric reasoning with 2D projection optimization, and introduces the pseudo-renderer, a differentiable module to approximate the true rendering operation, to synthesize novel depth maps for optimization.
This work proposes an approximate gradient for rasterization that enables the integration of rendering into neural networks and performs gradient-based 3D mesh editing operations, such as 2D-to-3D style transfer and 3D DeepDream, with 2D supervision for the first time.
This work introduces a novel method for the fast up-sampling of 3D objects in voxel space through networks that perform super-resolution on the six orthographic depth projections, and achieves state-of-the-art performance on 3D object reconstruction from RGB images on the ShapeNet dataset.
Novel, efficient 2D encodings for 3D geometry enable reconstructing full 3D shapes from a single image at high resolution, and clearly outperform previous octree-based approaches despite having a much simpler architecture using standard network components.
A novel framework for single-view and multi-view 3D object reconstruction, named Pix2Vox++ is proposed, which performs favorably against state-of-the-art methods in terms of both accuracy and efficiency.
Adding a benchmark result helps the community track progress.