3260 papers • 126 benchmarks • 313 datasets
Low resource named entity recognition is the task of using data and models available for one language for which ample such resources are available (e.g., English) to solve named entity recognition tasks in another, commonly more low-resource, language.
(Image credit: Papersgraph)
These leaderboards are used to track progress in low-resource-named-entity-recognition-11
Use these libraries to find low-resource-named-entity-recognition-11 models and implementations
No subtasks available.
Results show that using SMLM can achieve comparable results for Historic named entity recognition, even when they are only trained on contemporary texts, and it is shown that language model pre-training can be more effective than using transfer-learning with labeled datasets.
Evaluating on named entity recognition, it is shown that the proposed techniques for modulating the transfer are much more effective than strong baselines, including standard ensembling, and the unsupervised method rivals oracle selection of the single best individual model.
To the best of the knowledge, this approach is the first to leverage feature-dependent noise modeling with pre-initialized confusion matrices and improve upon other confusion-matrix based methods by up to 9%.
This paper proposes an unsupervised cross-lingual NER model that can transfer NER knowledge from one language to another in a completely unsuper supervised way without relying on any bilingual dictionary or parallel data.
This work proposes a method of “soft gazetteers” that incorporates ubiquitously available information from English knowledge bases, such as Wikipedia, into neural named entity recognition models through cross-lingual entity linking.
This work proposes a novel robust and domain-adaptive approach RDANER for low-resource NER, which only uses cheap and easily obtainable resources, and delivers competitive results against state-of-the-art methods which use difficultly obtainable domainspecific resources.
ANA, a tool to automatically annotate named entities in text based on entity lists, spans the whole pipeline from obtaining the lists to analyzing the errors of the distant supervision and it is shown that the F1-score can be increased by on average 18 points through distantly supervised data obtained by ANEA.
This work introduces an encoder evaluation framework, and uses it to systematically compare the performance of state-of-the-art pre-trained representations on the task of low-resource NER.
Adding a benchmark result helps the community track progress.