3260 papers • 126 benchmarks • 313 datasets
Lexical analysis is the process of converting a sequence of characters into a sequence of tokens (strings with an assigned and thus identified meaning). (Source: Adapted from Wikipedia)
(Image credit: Papersgraph)
These leaderboards are used to track progress in lexical-analysis-11
No benchmarks available.
Use these libraries to find lexical-analysis-11 models and implementations
A deep Bi-GRU-CRF network that jointly models word segmentation, part-of-speech tagging and named entity recognition tasks and achieves a 95.5% accuracy on the test set, roughly 13% relative error reduction over the best Chinese lexical analysis tool.
This paper fine-tune transformer-based language models pre-trained on several text corpora, some being general (E.g., Wikipedia, BooksCorpus), some being the corpora from which the CompLex Dataset was extracted, and others being from other specific domains such as Finance, Law, etc.
Based on the Aristotelian concept of potentiality vs. actuality allowing for the study of energy and dynamics in language, we propose a field approach to lexical analysis. Falling back on the distributional hypothesis to statistically model word meaning, we used evolving fields as a metaphor to express time-dependent changes in a vector space model by a combination of random indexing and evolving self-organizing maps (ESOM). To monitor semantic drifts within the observation period, an experiment was carried out on the term space of a collection of 12.8 million Amazon book reviews. For evaluation, the semantic consistency of ESOM term clusters was compared with their respective neighbourhoods in WordNet, and contrasted with distances among term vectors by random indexing. We found that at 0.05 level of significance, the terms in the clusters showed a high level of semantic consistency. Tracking the drift of distributional patterns in the term space across time periods, we found that consistency decreased, but not at a statistically significant level. Our method is highly scalable, with interpretations in philosophy.
A lexical analysis framework, the Pivot Analysis, is proposed, to quantitatively analyze the effects of these words in text attribute classification and transfer and identifies the future requirements and challenges of this task.
N-LTP, an open-source neural language technology platform supporting six fundamental Chinese NLP tasks: lexical analysis (Chinese word segmentation, part-of-speech tagging, and named entity recognition), syntactic parsing (dependency parsing), and semantic parsing (semantic dependency parsing and semantic role labeling).
The approach presented here is based on the principle that cultural affiliation can be inferred from the topics that people discuss among themselves, and uncovers a manifest North–South separation, and further contiguous and non-contiguous divisions that provide a comprehensive picture of modern American cultural areas.
Adding a benchmark result helps the community track progress.