3260 papers • 126 benchmarks • 313 datasets
"Camera relocalization, or image-based localization is a fundamental problem in robotics and computer vision. It refers to the process of determining camera pose from the visual scene representation and it is essential for many applications such as navigation of autonomous vehicles, structure from motion (SfM), augmented reality (AR) and simultaneous localization and mapping (SLAM)." (Source)
(Image credit: Papersgraph)
These leaderboards are used to track progress in camera-relocalization-10
No benchmarks available.
Use these libraries to find camera-relocalization-10 models and implementations
A multimodal camera relocalization framework that captures ambiguities and uncertainties with continuous mixture models defined on the manifold of camera poses, using Bingham distributions and a multivariate Gaussian to model the position, with an end-to-end deep neural network.
A Bayesian convolutional neural network is used to regress the 6-DOF camera pose from a single RGB image and an estimate of the model's relocalization uncertainty is obtained to improve state of the art localization accuracy on a large scale outdoor dataset.
A sample-balanced objective to encourage equal numbers of samples in the left and right sub-trees, and a novel backtracking scheme to remedy the incorrect 2D-3D correspondence predictions are proposed.
This work proposes an end-to-end framework, termed VMLoc, to fuse different sensor inputs into a common latent space through a variational Product-of-Experts (PoE) followed by attention-based fusion, and shows how camera localization can be accurately estimated through an unbiased objective function based on importance weighting.
KFNet extends the scene coordinate regression problem to the time domain in order to recursively establish 2D and 3D correspondences for the pose determination in temporal relocalization using a network architecture based on Kalman filtering in the context of Bayesian learning.
This paper adapts 3RScan - a recently introduced indoor RGB-D dataset designed for object instance re-localization - to create RIO10, a new long-term camera re- localization benchmark focused on indoor scenes and explores how state-of-the-art cameraRe-localizers perform according to these metrics.
This paper proposes a novel outlier-aware neural tree which bridges the two worlds, deep learning and decision tree approaches and outperforms the state-of-the-art approaches by around 30% on camera pose accuracy, while running comparably fast for evaluation.
This work introduces Deep Bingham Networks (DBN), a generic framework that can naturally handle pose-related uncertainties and ambiguities arising in almost all real life applications concerning 3D data, and proposed new training strategies so as to avoid mode or posterior collapse during training and to improve numerical stability.
This work designs a direct matching module that supplies a photometric supervision signal to refine the pose regression network via differentiable rendering and shows that the method can easily cope with additional unlabeled data without the need for external supervision.
Adding a benchmark result helps the community track progress.