3260 papers • 126 benchmarks • 313 datasets
XLM-R
(Image credit: Papersgraph)
These leaderboards are used to track progress in xlm-r-11
No benchmarks available.
Use these libraries to find xlm-r-11 models and implementations
No subtasks available.
It is shown that pretraining multilingual language models at scale leads to significant performance gains for a wide range of cross-lingual transfer tasks, and the possibility of multilingual modeling without sacrificing per-language performance is shown for the first time.
The MASSIVE dataset–Multilingual Amazon Slu resource package (SLURP) for Slot-filling, Intent classification, and Virtual assistant Evaluation is presented and modeling results on XLM-R and mT5 are presented, including exact match accuracy, intent classification accuracy, and slot- filling F1 score.
AdaptersHub is proposed, a framework that allows dynamic “stiching-in” of pre-trained adapters for different tasks and languages that enables scalable and easy access to sharing of task-specific models, particularly in low-resource scenarios.
Experimental outcomes indicate that XLM-R outdoes all other techniques by achieving the highest weighted f_1-score of 69.73% on the test data.
very close performances with CLIP on almost all tasks are obtained, suggesting that one can simply alter the text encoder in CLIP for extended capabilities such as multilingual understanding.
MAD-X is proposed, an adapter-based framework that enables high portability and parameter-efficient transfer to arbitrary tasks and languages by learning modular language and task representations and introduces a novel invertible adapter architecture and a strong baseline method for adapting a pretrained multilingual model to a new language.
BERTweet is presented, the first public large-scale pre-trained language model for English Tweets, trained using the RoBERTa pre-training procedure, producing better performance results than the previous state-of-the-art models on three Tweet NLP tasks.
This work introduces two powerful deep bidirectional transformer-based models, ARBERT and MARBERT, for a collection of diverse Arabic varieties, and introduces ARLUE, a new benchmark for multi-dialectal Arabic language understanding evaluation.
Adding a benchmark result helps the community track progress.