3260 papers • 126 benchmarks • 313 datasets
Sentence Compression is the task of reducing the length of text by removing non-essential content while preserving important facts and grammaticality.
(Image credit: Papersgraph)
These leaderboards are used to track progress in sentence-compression
Use these libraries to find sentence-compression models and implementations
This paper presents an N-best reranking method based on keyphrase extraction that significantly improves the informativity of the generated compressions.
A novel graph-based framework for abstractive meeting speech summarization that is fully unsupervised and does not rely on any annotations to take exterior semantic knowledge into account, and to design custom diversity and informativeness measures.
This work addresses the simplification problem with an encoder-decoder model coupled with a deep reinforcement learning framework, and explores the space of possible simplifications while learning to optimize a reward function that encourages outputs which are simple, fluent, and preserve the meaning of the input.
A globally normalized transition-based neural network model that achieves state-of-the-art part- of-speech tagging, dependency parsing and sentence compression results is introduced.
A way to automatically identify operations in a parallel corpus and introduce a sequence-labeling approach based on these annotations is devised, which provides insights on the types of transformations that different approaches can model.
A fully unsupervised, extractive text summarization system that leverages a submodularity framework that allows summaries to be generated in a greedy way while preserving near-optimal performance guarantees is presented.
This work develops a more comprehensive method to generate the story AMR graph using state-of-the-art co-reference resolution and Meta Nodes and outperforms the state of the art SAS method by 1.7% F1 score in node prediction.
The use of bilingual corpora which are abundantly available for training sentence compression models are advocated and a new parallel Multilingual Compression dataset is released which can be used to evaluate compression models across languages and genres.
This paper transforms the sequence to graph mapping problem to a word sequence to transition action sequence problem using a special transition system called a cache transition system, and presents a monotonic hard attention model for the transition framework to handle the strictly left-to-right alignment between each transition state and the current buffer input focus.
Although the models are underperform supervised models based on ROUGE scores, their models are competitive with a supervised baseline based on human evaluation for grammatical correctness and retention of meaning.
Adding a benchmark result helps the community track progress.