3260 papers • 126 benchmarks • 313 datasets
Paraphrase Generation involves transforming a natural language sentence to a new sentence, that has the same semantic meaning but a different syntactic or lexical surface form.
(Image credit: Papersgraph)
These leaderboards are used to track progress in paraphrase-generation-10
Use these libraries to find paraphrase-generation-10 models and implementations
No subtasks available.
The proposed method results in semantic embeddings and outperforms the state-of-the-art on the paraphrase generation and sentiment analysis task on standard datasets and is shown to be statistically significant.
This work proposes a latent bag of words (BOW) model for paraphrase generation that ground the semantics of a discrete latent variable by the target BOW to build a fully differentiable content planning and surface realization pipeline.
This work uses syntactic transformations to softly “reorder” the source sentence and guide the neural paraphrasing model, which retains the quality of the baseline approaches while giving a substantial increase in the diversity of the generated paraphrases.
Syntax Guided Controlled Paraphraser (SGCP), an end-to-end framework for syntactic paraphrase generation, is proposed and it is found that Sgcp can generate syntax-conforming sentences while not compromising on relevance.
This work is the first to explore deep learning models for paraphrase generation with a stacked residual LSTM network, where it adds residual connections between L STM layers for efficient training of deep LSTMs.
This work built and released the first evaluation corpus for Japanese paraphrase identification, which comprises 655 sentence pairs, and proposes a novel sentential paraphrase acquisition method that uses multiple machine translation systems and a monolingual corpus to extract negative candidates.
Quantitative evaluation of the proposed method on a benchmark paraphrase dataset demonstrates its efficacy, and its performance improvement over the state-of-the-art methods by a significant margin, whereas qualitative human evaluation indicate that the generated paraphrases are well-formed, grammatically correct, and are relevant to the input sentence.
This work introduces a novel model based on the encoder-decoder framework, called Word Embedding Attention Network (WEAN), which generates the words by querying distributed word representations (i.e. neural word embeddings), hoping to capturing the meaning of the according words.
The proposed model can transfer the prediction of lexico-semantic relations to a resource-lean target language without any training data and outperforms more complex state-of-the-art architectures on two benchmark datasets and exhibits stable performance across languages.
Adding a benchmark result helps the community track progress.