3260 papers • 126 benchmarks • 313 datasets
Summarization is the task of producing a shorter version of one or several documents that preserves most of the input's meaning.
(Image credit: Papersgraph)
These leaderboards are used to track progress in summarization-10
Use these libraries to find summarization-10 models and implementations
This work proposes several novel models that address critical problems in summarization that are not adequately modeled by the basic architecture, such as modeling key-words, capturing the hierarchy of sentence-to-word structure, and emitting words that are rare or unseen at training time.
MuLD is presented: a new long document benchmark consisting of only documents over 10,000 tokens, which requires models to successfully model long-term dependencies in the text and shows that models with increased context length are better able to solve the tasks presented.
A novel method to sparsify attention in the Transformer model by learning to select the most-informative token representations during the training process, thus focusing on the task-specific parts of an input.
Adding a benchmark result helps the community track progress.