3260 papers • 126 benchmarks • 313 datasets
( Image credit: One-Shot Learning for Semantic Segmentation )
(Image credit: Papersgraph)
These leaderboards are used to track progress in one-shot-segmentation-10
Use these libraries to find one-shot-segmentation-10 models and implementations
No subtasks available.
This work trains a network that, given a small set of annotated images, produces parameters for a Fully Convolutional Network (FCN), and uses this FCN to perform dense pixel-level prediction on a test image for the new semantic class.
This work proposes a system that can generate image segmentations based on arbitrary prompts at test time, and builds upon the CLIP model as a backbone which it extends with a transformer-based decoder that enables dense prediction.
A framework for the automatic one-shot segmentation of synthetic images generated by a StyleGAN that learns to segment synthetic images using a self-supervised contrastive clustering algorithm that projects the hidden features into a compact space for per-pixel classification.
An improved model that attends to multiple candidate locations, generates segmentation proposals to mask out background clutter and selects among the segmented objects is introduced, suggesting that such image recognition models based on an iterative refinement of object detection and foreground segmentation may provide a way to deal with highly cluttered scenes.
A framework to bridge the gap between information interaction between query and support images and to precisely localize the query objects is proposed, which outperforms other competitive methods and leads to a new state-of-the-art on both PASCAL VOC and MSCOCO dataset.
This work forms anatomy segmentation as a contour evolution process and model the evolution behavior by graph convolutional networks (GCNs) and demonstrates that the one-shot learning method significantly outperforms non-learning-based methods and performs competitively to the state-of-the-art fully supervised deep learning methods.
A novel Progressive One-shot Parsing network (POPNet) to address two critical challenges, i.e., testing bias and small sizes, which is based on two collaborative metric learning modules named Attention Guidance Module and Nearest Centroid Module.
This paper presents a local and multi-scale spatiotemporal representation learning method for image-to-image architectures trained on longitudinal images that exploits the spatiotsemporal self-similarity of learned multi- scale intra-subject image features for pretraining and develops several feature-wise regularizations that avoid degenerate representations.
An end-to-end human parsing framework that can quickly adapt to the novel classes and mitigate the testing bias issue is proposed, and a contrastive loss at the prototype level is added to enforce inter-class distances.
An innovative framework for 3D medical image segmentation with one-shot and weakly-supervised settings is presented and shows significant improvement over the state-of-the-art methods and performs robustly even under severe class imbalance and low contrast.
Adding a benchmark result helps the community track progress.