3260 papers • 126 benchmarks • 313 datasets
Spelling correction is the task of detecting and correcting spelling mistakes.
(Image credit: Papersgraph)
These leaderboards are used to track progress in spelling-correction-5
No benchmarks available.
Use these libraries to find spelling-correction-5 models and implementations
No subtasks available.
An approach to training neural networks to generate sequences using actor-critic methods from reinforcement learning (RL) that condition the critic network on the ground-truth output, and shows that this method leads to improved performance on both a synthetic task, and for German-English machine translation.
MoNoise is a normalization model focused on generalizability and efficiency, it aims at being easily reusable and adaptable, based on a modular candidate generation in which each module is responsible for a different type of normalization action.
This work proposes robust to noise word embeddings model, which outperforms existing commonly used models, like fasttext and word2vec in different tasks, and investigates the noise robustness of current models in different natural language processing tasks.
This work identifies three key ingredients of high-quality tokenization repair, all missing from previous work: deep language models with a bidirectional component, training the models on text with spelling errors, and making use of the space information already present.
A novel approach based on a systematic, recursive exploration of skip n-gram models which are interpolated using modified Kneser-Ney smoothing is introduced which generalizes language models as it contains the classical interpolation with lower order models as a special case.
Inspired by the findings from the Cmabrigde Uinervtisy effect, a word recognition model based on a semi-character level recurrent neural network (scRNN) is proposed that has significantly more robust performance in word spelling correction compared to existing spelling checkers and character-based convolutional neural network.
The approaches and results on the WAT 2016 shared translation tasks were described, which tried to use both an example-based machine translation (MT) system and a neural MT system.
This work adapts machine translation to grammatical error correction, identifying how components of the statistical MT pipeline can be modified for this task and analyzing how each modification impacts system performance.
An unsupervised context-sensitive spelling correction method for clinical free-text that uses word and character n-gram embeddings that generates misspelling replacement candidates and ranks them according to their semantic fit, by calculating a weighted cosine similarity between the vectorized representation of a candidate and the misspelling context.
Adding a benchmark result helps the community track progress.