3260 papers • 126 benchmarks • 313 datasets
A classic problem in natural-language generation (NLG) involves taking structured data, such as a table, as input, and producing text that adequately and fluently describes this data as output. Unlike machine translation, which aims for complete transduction of the sentence to be translated, this form of NLG is usually taken to require addressing (at least) two separate challenges: what to say, the selection of an appropriate subset of the input data to discuss, and how to say it, the surface realization of a generation. ( Image credit: Data-to-Text Generation with Content Selection and Planning )
(Image credit: Papersgraph)
These leaderboards are used to track progress in data-to-text-generation-5
Use these libraries to find data-to-text-generation-5 models and implementations
A new, large-scale corpus of data records paired with descriptive documents is introduced, a series of extractive evaluation methods for analyzing performance are proposed, and baseline results are obtained using current neural generation methods.
This paper investigates strategies to encode system and reference texts to devise a metric that shows a high correlation with human judgment of text quality and validate the new metric, namely MoverScore, on a number of text generation tasks.
It is suggested that the PLMs benefit from similar facts seen during pretraining or fine-tuning, such that they perform well even when the input graph is reduced to a simple bag of node and edge labels.
The E2E dataset poses new challenges: (1) its human reference texts show more lexical richness and syntactic variation, including discourse phenomena; (2) generating from this set requires content selection, which promises more natural, varied and less template-like system utterances.
This work presents a simple and effective oversampling method based on k-means clustering and SMOTE (synthetic minority oversampled technique), which avoids the generation of noise and effectively overcomes imbalances between and within classes.
This work presents a neural network architecture which incorporates content selection and planning without sacrificing end-to-end training and shows that this model outperforms strong baselines improving the state-of-the-art on the recently released RotoWire dataset.
This paper proposes an alternative encoder based on graph convolutional networks that directly exploits the input structure and reports results on two graph-to-sequence datasets that empirically show the benefits of explicitly encoding the input graph structure.
It is shown that rare items strongly impact performance and that combining delexicalisation and copying yields the strongest improvement; that copying underperforms for rare and unseen items and that the impact of these two mechanisms greatly varies depending on how the dataset is constructed and on how it is split into train, dev and test.
This work considers two pragmatic modeling methods for text generation: one where pragmatics is imposed by information preservation, and another where prag matics isimposed by explicit modeling of distractors.
Adding a benchmark result helps the community track progress.