3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in cross-lingual-ner-4
Use these libraries to find cross-lingual-ner-4 models and implementations
No subtasks available.
This paper shows that a standard Transformer architecture can be used with minimal modifications to process byte sequences, characterize the trade-offs in terms of parameter count, training FLOPs, and inference speed, and shows that byte-level models are competitive with their token-level counterparts.
It is experimentally demonstrated that high capacity multilingual language models applied in a zero-shot (model-based cross-lingual transfer) setting consistently outperform data-basedCross-lingUAL transfer approaches.
This paper explores the broader cross-lingual potential of mBERT (multilingual) as a zero shot language transfer model on 5 NLP tasks covering a total of 39 languages from various language families: NLI, document classification, NER, POS tagging, and dependency parsing.
A simple and novel framework is proposed that combines two previously mutually-exclusive approaches to learning unified multilingual representations using monolingual and cross-lingual objectives jointly, and outperforms existing methods on the MUSE bilingual lexicon induction (BLI) benchmark.
The analysis shows that larger output embeddings prevent the model's last layers from overspecializing to the pre-training task and encourage Transformer representations to be more general and more transferable to other tasks and languages.
T-Projection is presented, a novel approach for annotation projection that leverages large pretrained text-to-text language models and state-of-the-art machine translation technology and can help to automatically alleviate the lack of high-quality training data for sequence labeling tasks.
This model leverages adversarial networks to learn language-invariant features, and mixture-of-experts models to dynamically exploit the similarity between the target language and each individual source language to further boost target language performance.
This work proposes a system that improves over prior entity-projection methods by leveraging machine translation systems twice: first for translating sentences and subsequently for translating entities; and matching entities based on orthographic and phonetic similarity; and identifying matches based on distributional statistics derived from the dataset.
This paper presents a meta-learning algorithm to find a good model parameter initialization that could fast adapt to the given test case and proposes to construct multiple pseudo-NER tasks for meta-training by computing sentence similarities.
This paper proposes an unsupervised cross-lingual NER model that can transfer NER knowledge from one language to another in a completely unsuper supervised way without relying on any bilingual dictionary or parallel data.
Adding a benchmark result helps the community track progress.