3260 papers • 126 benchmarks • 313 datasets
Sentence Compression is the task of reducing the length of text by removing non-essential content while preserving important facts and grammaticality.
(Image credit: Papersgraph)
These leaderboards are used to track progress in sentence-compression-5
Use these libraries to find sentence-compression-5 models and implementations
A novel graph-based framework for abstractive meeting speech summarization that is fully unsupervised and does not rely on any annotations to take exterior semantic knowledge into account, and to design custom diversity and informativeness measures.
A globally normalized transition-based neural network model that achieves state-of-the-art part- of-speech tagging, dependency parsing and sentence compression results is introduced.
This work addresses the simplification problem with an encoder-decoder model coupled with a deep reinforcement learning framework, and explores the space of possible simplifications while learning to optimize a reward function that encourages outputs which are simple, fluent, and preserve the meaning of the input.
A way to automatically identify operations in a parallel corpus and introduce a sequence-labeling approach based on these annotations is devised, which provides insights on the types of transformations that different approaches can model.
A fully unsupervised, extractive text summarization system that leverages a submodularity framework that allows summaries to be generated in a greedy way while preserving near-optimal performance guarantees is presented.
This paper transforms the sequence to graph mapping problem to a word sequence to transition action sequence problem using a special transition system called a cache transition system, and presents a monotonic hard attention model for the transition framework to handle the strictly left-to-right alignment between each transition state and the current buffer input focus.
This work develops a more comprehensive method to generate the story AMR graph using state-of-the-art co-reference resolution and Meta Nodes and outperforms the state of the art SAS method by 1.7% F1 score in node prediction.
Although the models are underperform supervised models based on ROUGE scores, their models are competitive with a supervised baseline based on human evaluation for grammatical correctness and retention of meaning.
The use of bilingual corpora which are abundantly available for training sentence compression models are advocated and a new parallel Multilingual Compression dataset is released which can be used to evaluate compression models across languages and genres.
Adding a benchmark result helps the community track progress.