3260 papers • 126 benchmarks • 313 datasets
Co-Salient Object Detection is a computational problem that aims at highlighting the common and salient foreground regions (or objects) in an image group. Please also refer to the online benchmark: http://dpfan.net/cosod3k/ ( Image credit: Taking a Deeper Look at Co-Salient Object Detection, CVPR2020 )
(Image credit: Papersgraph)
These leaderboards are used to track progress in co-salient-object-detection-21
Use these libraries to find co-salient-object-detection-21 models and implementations
No subtasks available.
This paper presents an edge guidance network (EGNet) for salient object detection with three steps to simultaneously model these two kinds of complementary information in a single network to solve the complementarity between salient edge information and salient object information.
In this paper, we address a new task called instance co-segmentation. Given a set of images jointly covering object instances of a specific category, instance co-segmentation aims to identify all of these instances and segment each of them, i.e. generating one mask for each instance. This task is important since instance-level segmentation is preferable for humans and many vision applications. It is also challenging because no pixel-wise annotated training data are available and the number of instances in each image is unknown. We solve this task by dividing it into two sub-tasks, co-peak search and instance mask segmentation. In the former sub-task, we develop a CNN-based network to detect the co-peaks as well as co-saliency maps for a pair of images. A co-peak has two endpoints, one in each image, that are local maxima in the response maps and similar to each other. Thereby, the two endpoints are potentially covered by a pair of instances of the same category. In the latter subtask, we design a ranking function that takes the detected co-peaks and co-saliency maps as inputs and can select the object proposals to produce the final results. Our method for instance co-segmentation and its variant for object colocalization are evaluated on four datasets, and achieve favorable performance against the state-of-the-art methods. The source codes and the collected datasets are available at https://github.com/KuangJuiHsu/DeepCO3/
A novel CoEG-Net is proposed that augments the authors' prior model EGNet with a co-attention projection strategy to enable fast common information learning and fully leverages previous large-scale SOD datasets and significantly improves the model scalability and stability.
A gradient-induced co-saliency detection (GICD) method that first abstracts a consensus representation for the grouped images in the embedding space; then, by comparing the single image with consensus representation, utilizes the feedback gradient information to induce more attention to the discriminative co- salient features.
This work presents a novel adaptive graph convolutional network with attention graph clustering (GCAGC), and proposes a unified framework with encoder-decoder structure to jointly train and optimize the graph convolutionsal network, attention graph cluster, and co-saliency detection decoder in an end-to-end manner.
This paper presents an end-to-end collaborative aggregation-and-distribution network (CoADNet) to capture both salient and repetitive visual patterns from multiple images, and develops a group consistency preserving decoder tailored for the CoSOD task.
Co-salient object detection (CoSOD) is a newly emerging and rapidly growing branch of salient object detection (SOD), which aims to detect the co-occurring salient objects in multiple images. However, existing CoSOD datasets often have a serious data bias, which assumes that each group of images contains salient objects of similar visual appearances. This bias results in the ideal settings and the effectiveness of the models, trained on existing datasets, may be impaired in real-life situations, where the similarity is usually semantic or conceptual. To tackle this issue, we first collect a new high-quality dataset, named CoSOD3k, which contains 3,316 images divided into 160 groups with multiple level annotations, i.e., category, bounding box, object, and instance levels. CoSOD3k makes a significant leap in terms of diversity, difficulty and scalability, benefiting related vision tasks. Besides, we comprehensively summarize 34 cutting-edge algorithms, benchmarking 19 of them over four existing CoSOD datasets (MSRC, iCoSeg, Image Pair and CoSal2015) and our CoSOD3k with a total of ∼61K images (largest scale), and reporting group-level performance analysis. Finally, we discuss the challenge and future work of CoSOD. Our study would give a strong boost to growth in the CoSOD community. Benchmark toolbox and results are available on our project page.
This paper’s new technical contributions on a number of important downstream computer vision applications including content aware co-segmentation, co-localization based automatic thumbnails, etc are demonstrated.
The very first blackbox joint adversarial exposure and noise attack (Jadena), where they jointly and locally tune the exposure and additive perturbations of the image according to a newly designed high-feature-level contrast-sensitive loss function, leads to significant performance degradation on various co- saliency detection datasets and makes the co-salient objects undetectable.
A novel consensus-aware dynamic convolution model is proposed to explicitly and effectively perform the "summarize and search" process to summarize consensus image features and generate dynamic kernels from consensus features to encode the summarized consensus knowledge.
Adding a benchmark result helps the community track progress.