3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in pretrained-multilingual-language-models-11
No benchmarks available.
Use these libraries to find pretrained-multilingual-language-models-11 models and implementations
No subtasks available.
It is found that replacing the original multilingual tokenizer with the specialized monolingual tokenizer improves the downstream performance of the multilingual model for almost every task and language.
The experiments show that the MWP solvers may not be transferred to a different language even if the target expressions share the same numerical constants and operator set, and it can be better generalized if problem types exist on both source language and target language.
These evaluations on part-of-speech tagging, universal dependency parsing, and named entity recognition in nine diverse low-resource languages uphold the viability of these approaches while raising new questions around how to optimally adapt multilingual models to low- resource settings.
This work proposes a robust and effective two-stage contrastive learning framework for the BLI task, and proposes to refine standard cross-lingual linear maps between static word embeddings (WEs) via a Contrastive learning objective.
In this work, multifaceted investigations on fine-tuning and adapters for summarization tasks with varying complexity: language, domain, and task transfer provide insights on multilinguality, model convergence, and robustness, hoping to shed light on the pragmatic choice of fine- tuning or adapters in abstractive summarization.
After investigating several ways to boost the robustness of multilingual models in this setting, this work proposes Robust Contrastive Pretraining (RCP), which combines data augmentation with a contrastive loss term at the pretraining stage and achieves large improvements on noisy data.
This work evaluates three multilingual models on MozArt –mBERT, XLM-R, and mT5– and shows that across the four target languages, the three models exhibit different levels of group disparity, e.g., exhibiting near-equal risk for Spanish, but high levels of disparity for German.
An effective method to train multilingual IR systems when only English IR training data and some parallel corpora between English and other languages are available is presented and a semantic contrastive loss to align representations of parallel sentences that share the same semantics in different languages is designed.
This work proposes a novel semi-supervised post-hoc reranking method termed BLICEr (BLI with Cross-Encoder Reranking), applicable to any precalculated CLWE space, which improves their BLI capability and substantially outperforms a series of strong baselines across the board.
Adding a benchmark result helps the community track progress.