3260 papers • 126 benchmarks • 313 datasets
Camouflaged object segmentation (COS) or Camouflaged object detection (COD), which was originally promoted by T.-N. Le et al. (2017), aims to identify objects that conceal their texture into the surrounding environment. The high intrinsic similarities between the target object and the background make COS/COD far more challenging than the traditional object segmentation task. Also, refer to the online benchmarks on CAMO dataset, COD dataset, and online demo. ( Image source: Anabranch Network for Camouflaged Object Segmentation )
(Image credit: Papersgraph)
These leaderboards are used to track progress in camouflaged-object-segmentation-2
Use these libraries to find camouflaged-object-segmentation-2 models and implementations
This paper presents UNet++, a new, more powerful architecture for medical image segmentation where the encoder and decoder sub-networks are connected through a series of nested, dense skip pathways, and argues that the optimizer would deal with an easier learning task when the feature maps from the decoder and encoder networks are semantically similar.
This paper presents an edge guidance network (EGNet) for salient object detection with three steps to simultaneously model these two kinds of complementary information in a single network to solve the complementarity between salient edge information and salient object information.
The proposed Boundary-Aware Segmentation Network (BASNet), which comprises a predict-refine architecture and a hybrid loss, for highly accurate image segmentation, and further developed two (close to) commercial applications: AR COPY&PASTE, in which BASNet is integrated with augmented reality, and OBJECT CUT, which is a web-based tool for automatic object background removal.
The F3Net is able to segment salient object regions accurately and provide clear local details and outperforms state-of-the-art approaches on six evaluation metrics.
Quantitative and qualitative evaluations on five challenging datasets across six metrics show that the PraNet improves the segmentation accuracy significantly, and presents a number of advantages in terms of generalizability, and real-time segmentation efficiency.
Deep Convolutional Neural Networks have been adopted for salient object detection and achieved the state-of-the-art performance. Most of the previous works however focus on region accuracy but not on the boundary quality. In this paper, we propose a predict-refine architecture, BASNet, and a new hybrid loss for Boundary-Aware Salient object detection. Specifically, the architecture is composed of a densely supervised Encoder-Decoder network and a residual refinement module, which are respectively in charge of saliency prediction and saliency map refinement. The hybrid loss guides the network to learn the transformation between the input image and the ground truth in a three-level hierarchy -- pixel-, patch- and map- level -- by fusing Binary Cross Entropy (BCE), Structural SIMilarity (SSIM) and Intersection-over-Union (IoU) losses. Equipped with the hybrid loss, the proposed predict-refine architecture is able to effectively segment the salient object regions and accurately predict the fine structures with clear boundaries. Experimental results on six public datasets show that our method outperforms the state-of-the-art methods both in terms of regional and boundary evaluation measures. Our method runs at over 25 fps on a single GPU. The code is available at: https://github.com/NathanUA/BASNet.
We present a comprehensive study on a new task named camouflaged object detection (COD), which aims to identify objects that are “seamlessly” embedded in their surroundings. The high intrinsic similarities between the target object and the background make COD far more challenging than the traditional object detection task. To address this issue, we elaborately collect a novel dataset, called COD10K, which comprises 10,000 images covering camouflaged objects in various natural scenes, over 78 object categories. All the images are densely annotated with category, bounding-box, object-/instance-level, and matting-level labels. This dataset could serve as a catalyst for progressing many vision tasks, e.g., localization, segmentation, and alpha-matting, etc. In addition, we develop a simple but effective framework for COD, termed Search Identification Network (SINet). Without any bells and whistles, SINet outperforms various state-of-the-art object detection baselines on all datasets tested, making it a robust, general framework that can help facilitate future research in COD. Finally, we conduct a large-scale COD study, evaluating 13 cutting-edge models, providing some interesting findings, and showing several potential applications. Our research offers the community an opportunity to explore more in this new field. The code will be available at https://github.com/DengPingFan/SINet/.
This work introduces a latent variable model within the transformer framework, termed the inferential generative adversarial network (iGAN), and applies it to fully supervised salient object detection, explaining that iGAN within the transformer framework leads to both accurate and reliable salient object detection.
This paper proposes a general end-to-end network, called the Anabranch Network, that leverages both classification and segmentation tasks and possesses the second branch for classification to predict the probability of containing camouflaged object(s) in an image.
A novel Context-aware Cross-level Fusion Network (C2F-Net) to address the challenging COD task and proposes an Attention-induced Cross- level Fusion Module (ACFM) to integrate the multi-level features with informative attention coefficients.
Adding a benchmark result helps the community track progress.