3260 papers • 126 benchmarks • 313 datasets
Optical flow is a two-dimensional motion field in the image plane. It is the projection of the three-dimensional motion of the world. If the world is completely non-rigid, the motions of the points in the scene may all be indepen- dent of each other. One representation of the scene motion is therefore a dense three-dimensional vector field defined for every point on every surface in the scene. By analogy with optical flow, we refer to this three-dimensional motion field as scene flow. Source: Vedula, Sundar, et al. "Three-dimensional scene flow." IEEE transactions on pattern analysis and machine intelligence 27.3 (2005): 475-480. pdf
(Image credit: Papersgraph)
These leaderboards are used to track progress in scene-flow-estimation-5
Use these libraries to find scene-flow-estimation-5 models and implementations
This work introduces a new large-scale dataset for scene flow estimation derived from corresponding tracked 3D objects, which is 1,000 times larger than previous real-world datasets in terms of the number of annotated frames, and designs human-interpretable metrics that better capture real world aspects by accounting for ego-motion and providing breakdowns per object type.
This paper proposes three synthetic stereo video datasets with sufficient realism, variation, and size to successfully train large networks and presents a convolutional network for real-time disparity estimation that provides state-of-the-art results.
A novel deep neural network architecture for end-to-end scene flow estimation that directly operates on large-scale 3D point clouds is presented and shows great generalization ability on real-world data and on different point densities without fine-tuning.
This work proposes a novel neural network architecture called MeteorNet for learning representations for dynamic 3D point cloud sequences that shows stronger performance than previous grid-based methods while achieving state-of-the-art performance on Synthia.
It is shown that recent OT solvers improve both optimization-based and deep learning methods for point cloud registration, boosting accuracy at an affordable computational cost and providing a new key method for the computer vision toolbox.
This work aims to address the above challenges and estimate scene flow from 4-D radar point clouds by leveraging self-supervised learning and a robust scene flow estimation architecture and three novel losses are bespoken designed to cope with intractable radar data.
Using a novel evaluation methodology based on a super-resolved UHD ground truth, the Spring benchmark can assess the quality of fine structures and provides further detailed performance statistics on different image regions.
This paper proposes to learn the rigidity of a scene in a supervised manner from a large collection of dynamic scene data, and directly infer a rigidity mask from two sequential images with depths, and shows how the proposed framework outperforms current state-of-the-art scene flow estimation methods in challenging dynamic scenes.
This paper introduces DeFlow which enables a transition from voxel-based features to point features using Gated Recurrent Unit (GRU) refinement, and formulate a novel loss function that accounts for the data imbalance between static and dynamic points.
This paper introduces a first effort to apply a deep learning method for direct estimation of scene flow by presenting a fully convolutional neural network with an encoder-decoder (ED) architecture.
Adding a benchmark result helps the community track progress.