3260 papers • 126 benchmarks • 313 datasets
In generalized zero shot learning (GZSL), the set of classes are split into seen and unseen classes, where training relies on the semantic features of the seen and unseen classes and the visual representations of only the seen classes, while testing uses the visual representations of the seen and unseen classes.
(Image credit: Papersgraph)
These leaderboards are used to track progress in generalized-zero-shot-learning-5
Use these libraries to find generalized-zero-shot-learning-5 models and implementations
No subtasks available.
It is found that non-linear aggregators such as LSTMs or transformers lead to significant improvements on zero-shot tasks, and ZSL-KG outperforms the best performing graph neural networks with linear aggregators by an average of 3.8 points of accuracy.
This work gives a theoretical explanation to two popular tricks used in zero-shot learning: normalize+scale and attributes normalization and proposes Class Normalization (CN): a normalization scheme, which alleviates this issue both provably and in practice.
A novel generative adversarial network (GAN) that synthesizes CNN features conditioned on class-level semantic information, offering a shortcut directly from a semantic descriptor of a class to a class-conditional feature distribution.
The proposed hybrid GZSL framework with contrastive embedding, named CE-GZSL, is evaluated, and the results show that the CEGZSL method can outperform the state-of-the-arts by a significant margin on three datasets.
This work proposes a model where a shared latent space of image features and class embeddings is learned by modality-specific aligned variational autoencoders, and align the distributions learned from images and from side-information to construct latent features that contain the essential multi-modal information associated with unseen classes.
This paper proposes a boundary based Out-of-Distribution (OOD) classifier which classifies the unseen and seen domains by only using seen samples for training and extensively validate the approach on five popular benchmark datasets including AWA1, AWA2, CUB, FLO and SUN.
A study aimed on evaluating the adversarial robustness of ZSL and GZSL models by leveraging the well-established label embedding model and subjecting it to a set of established adversarial attacks and defenses across multiple datasets.
It is shown that there is a large gap between the performance of existing approaches and the performance limit of GZSL, suggesting that improving the quality of class semantic embeddings is vital to improving ZSL.
This paper proposes the use of a constraint based on a new regularization for the GAN training that forces the generated visual features to reconstruct their original semantic features in a multi-modal cycle-consistent manner.
This work learns to map domain knowledge about novel “unseen” classes onto this dictionary of learned concepts and optimizes for network parameters that can effectively combine these concepts – essentially learning classifiers by discovering and composing learned semantic concepts in deep networks.
Adding a benchmark result helps the community track progress.