3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in zero-shot-cross-lingual-transfer-8
Use these libraries to find zero-shot-cross-lingual-transfer-8 models and implementations
No subtasks available.
An architecture to learn joint multilingual sentence representations for 93 languages, belonging to more than 30 different families and written in 28 different scripts using a single BiLSTM encoder with a shared byte-pair encoding vocabulary for all languages, coupled with an auxiliary decoder and trained on publicly available parallel corpora.
This paper explores the broader cross-lingual potential of mBERT (multilingual) as a zero shot language transfer model on 5 NLP tasks covering a total of 39 languages from various language families: NLI, document classification, NER, POS tagging, and dependency parsing.
This paper introduces ELECTRA-style tasks to cross-lingual language model pre-training, and pretrain the model, named as XLM-E, on both multilingual and parallel corpora, which outperforms the baseline models on various cross-lingsual understanding tasks with much less computation cost.
This paper extends previous work on zero-shot cross-lingual transfer learning by fine-tuning a multilingually pretrained wav2vec 2.0 model to transcribe unseen languages by mapping phonemes of the training languages to the target language using articulatory features.
This work introduces a new fine-tuning method that outperforms adapters in zero-shot cross-lingual transfer by a large margin in a series of multilingual benchmarks, including Universal Dependencies, MasakhaNER, and AmericasNLI.
This work proposes a novel approach to specializing the full distributional vocabulary by combining a standard L2-distance loss with a adversarial loss, and proposes a cross-lingual transfer method for zero-shot specialization which successfully specializes a full target distributional space without any lexical knowledge in the target language and without any bilingual data.
This paper proposes Cross-Lingual BERT Transformation (CLBT), a simple and efficient approach to generate cross-lingual contextualized word embeddings based on publicly available pre-trained BERT models (Devlin et al., 2018).
Experimental results on question generation and abstractive summarization show that the model outperforms the machine-translation-based pipeline methods for zero-shot cross-lingual generation and improves NLG performance of low-resource languages by leveraging rich-resource language data.
This work considers the setting of training models on multiple different languages at the same time, when little or no data is available for languages other than English, and demonstrates the consistent effectiveness of meta-learning for a total of 15 languages.
A Bayesian generative model for the space of neural parameters is proposed that can be factorized into latent variables for each language and each task, and infer the posteriors over such latent variables based on data from seen task–language combinations through variational inference.
Adding a benchmark result helps the community track progress.