3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in cross-lingual-ner-7
Use these libraries to find cross-lingual-ner-7 models and implementations
No subtasks available.
This paper shows that a standard Transformer architecture can be used with minimal modifications to process byte sequences, characterize the trade-offs in terms of parameter count, training FLOPs, and inference speed, and shows that byte-level models are competitive with their token-level counterparts.
It is experimentally demonstrated that high capacity multilingual language models applied in a zero-shot (model-based cross-lingual transfer) setting consistently outperform data-basedCross-lingUAL transfer approaches.
This paper explores the broader cross-lingual potential of mBERT (multilingual) as a zero shot language transfer model on 5 NLP tasks covering a total of 39 languages from various language families: NLI, document classification, NER, POS tagging, and dependency parsing.
The analysis shows that larger output embeddings prevent the model's last layers from overspecializing to the pre-training task and encourage Transformer representations to be more general and more transferable to other tasks and languages.
A simple and novel framework is proposed that combines two previously mutually-exclusive approaches to learning unified multilingual representations using monolingual and cross-lingual objectives jointly, and outperforms existing methods on the MUSE bilingual lexicon induction (BLI) benchmark.
T-Projection is presented, a novel approach for annotation projection that leverages large pretrained text-to-text language models and state-of-the-art machine translation technology and can help to automatically alleviate the lack of high-quality training data for sequence labeling tasks.
This model leverages adversarial networks to learn language-invariant features, and mixture-of-experts models to dynamically exploit the similarity between the target language and each individual source language to further boost target language performance.
This work proposes a system that improves over prior entity-projection methods by leveraging machine translation systems twice: first for translating sentences and subsequently for translating entities; and matching entities based on orthographic and phonetic similarity; and identifying matches based on distributional statistics derived from the dataset.
This paper presents a meta-learning algorithm to find a good model parameter initialization that could fast adapt to the given test case and proposes to construct multiple pseudo-NER tasks for meta-training by computing sentence similarities.
This paper proposes an unsupervised cross-lingual NER model that can transfer NER knowledge from one language to another in a completely unsuper supervised way without relying on any bilingual dictionary or parallel data.
Adding a benchmark result helps the community track progress.