3260 papers • 126 benchmarks • 313 datasets
Extreme Multi-Label Classification is a supervised learning problem where an instance may be associated with multiple labels. The two main problems are the unbalanced labels in the dataset and the amount of different labels.
(Image credit: Papersgraph)
These leaderboards are used to track progress in extreme-multi-label-classification-41
No benchmarks available.
Use these libraries to find extreme-multi-label-classification-41 models and implementations
No subtasks available.
A new self-supervised learning framework, Graph Information Aided Node feature exTraction (GIANT), which makes use of the eXtreme Multi-label Classification (XMC) formalism, which is crucial for fine-tuning the language model based on graph information, and scales to large datasets.
X-Transformer is proposed, the first scalable approach to fine-tuning deep transformer models for the XMC problem and achieves new state-of-the-art results on four XMC benchmark datasets.
This paper thoroughly investigates probabilistic label trees (PLTs) which can be treated as a generalization of hierarchical softmax for multi-label problems and proves the consistency of PLTs for a wide spectrum of performance metrics.
This paper analyzes generalized metrics budgeted at k in the expected test utility (ETU) framework and derives optimal prediction rules and construct computationally efficient approximations with provable regret guarantees and robustness against model misspecification.
This work proposes a general program, $\texttt{Infer--Retrieve--Rank}$, that defines multi-step interactions between LMs and retrievers to efficiently tackle multi-label classification problems and attains state-of-the-art results across three benchmarks.
This paper proposes a practical deep embedding method for extreme multi-label classification, which harvests the ideas of non-linear embedding and graph priors-based label space modeling simultaneously.
A novel approach to XMLC termed the Sparse Weighted Nearest-Neighbor Method, which can be derived as a fast implementation of state-of-the-art (SOTA) one-versus-rest linear classifiers for very sparse datasets with superior performance to the SOTA models on a dataset with 3 million labels.
This work poses the learning task in extreme classification with large number of tail-labels as learning in the presence of adversarial perturbations to motivate a robust optimization framework and equivalence to a corresponding regularized objective.
Adding a benchmark result helps the community track progress.