3260 papers • 126 benchmarks • 313 datasets
RGB-D Salient object detection (SOD) aims at distinguishing the most visually distinctive objects or regions in a scene from the given RGB and Depth data. It has a wide range of applications, including video/image segmentation, object recognition, visual tracking, foreground maps evaluation, image retrieval, content-aware image editing, information discovery, photosynthesis, and weakly supervised semantic segmentation. Here, depth information plays an important complementary role in finding salient objects. Online benchmark: http://dpfan.net/d3netbenchmark. ( Image credit: Rethinking RGB-D Salient Object Detection: Models, Data Sets, and Large-Scale Benchmarks, TNNLS20 )
(Image credit: Papersgraph)
These leaderboards are used to track progress in rgb-d-salient-object-detection
Use these libraries to find rgb-d-salient-object-detection models and implementations
No subtasks available.
This paper provides a comprehensive survey of RGB-D based salient object detection models from various perspectives, and review related benchmark datasets in detail, and investigates the ability of existing models to detect salient objects.
Qualitative and quantitative results on six challengingRGB-D benchmark datasets show the first stochastic framework to employ uncertainty for RGB-D saliency detection by learning from the data labeling process superior performance in learning the distribution of saliency maps.
A convolutional neural network (CNN) model, named CIR-Net, based on the novel cross-modality interaction and refinement is presented, which outperforms the state-of-the-art saliency detectors both qualitatively and quantitatively.
A new architecture is presented—PDNet, a robust prior-model guided depth-enhanced network for RGB-D salient object detection, composed of a master network for processing RGB values, and a sub-network making full use of depth cues and incorporate depth-based features into the master network.
The large availability of depth sensors provides valuable complementary information for salient object detection (SOD) in RGBD images. However, due to the inherent difference between RGB and depth information, extracting features from the depth channel using ImageNet pre-trained backbone models and fusing them with RGB features directly are sub-optimal. In this paper, we utilize contrast prior, which used to be a dominant cue in none deep learning based SOD approaches, into CNNs-based architecture to enhance the depth information. The enhanced depth cues are further integrated with RGB features for SOD, using a novel fluid pyramid integration, which can make better use of multi-scale cross-modal features. Comprehensive experiments on 5 challenging benchmark datasets demonstrate the superiority of the architecture CPFP over 9 state-of-the-art alternative methods.
It is demonstrated that D3Net can be used to efficiently extract salient object masks from real scenes, enabling effective background-changing application with a speed of 65 frames/s on a single GPU.
This paper uses the inherent multi-modal and multi-level nature of RGB-D salient object detection to devise a novel Bifurcated Backbone Strategy Network (BBS-Net), which significantly outperforms 18 state-of-the-art (SOTA) models on eight challenging datasets under five evaluation measures.
This paper proposes a novel Cross-Modal Weighting (CMW) strategy to encourage comprehensive interactions between RGB and depth channels for RGB-D SOD, and designs a composite loss function that summarizes the errors between intermediate predictions and ground truth over different scales.
A novel collaborative learning framework where edge, depth and saliency are leveraged in a more efficient way, which solves those problems tactfully and makes the model more lightweight, faster and more versatile.
The proposed JL-DCF module provides robust saliency feature learning by exploiting cross-modal commonality via a Siamese network, while the DCF module is introduced for complementary feature discovery.
Adding a benchmark result helps the community track progress.