3260 papers • 126 benchmarks • 313 datasets
Text Infilling is the task of predicting missing spans of text which are consistent with the preceding and subsequent text. Text Infilling is a generalization of the cloze task—cloze historically refers to infilling individual words. Source: Enabling Language Models to Fill in the Blanks
(Image credit: Papersgraph)
These leaderboards are used to track progress in text-infilling-10
No benchmarks available.
Use these libraries to find text-infilling-10 models and implementations
No subtasks available.
It is shown that humans have difficulty identifying sentences infilled by the approach, which can enable LMs to infill entire sentences effectively on three different domains: short stories, scientific abstracts, and lyrics.
A story-centric benchmark named LOT is proposed for evaluating Chinese long text modeling, which aggregates two understanding tasks and two generation tasks and shows that LongLM outperforms similar-sized pretraining models substantially on both the understanding and generation tasks in LOT.
This work introduces a hybrid generation approach inspired by traditional concept-to-text systems for generating comparative summaries that leads to more faithful, relevant and aggregation-sensitive summarization -- while being equally fluent.
A novel method for measurably adjusting the semantics of text while preserving its sentiment and fluency, a task the authors call semantic text exchange is presented and a pipeline called SMERTI is introduced that combines entity replacement, similarity masking, and text infilling.
This paper proposes DeLorean, a new unsupervised decoding algorithm that can flexibly incorporate both the past and future contexts using only off-the-shelf, left-to-right language models and no supervision.
An iterative inference algorithm based on gradient search is proposed, which could be the first inference algorithm that can be broadly applied to any neural sequence generative models for text infilling tasks.
This paper conducts extensive experiments by using SSR to improve the typical Seq2Seq pre-trained model T5 in a continual pre-training setting and show substantial improvements over T5 on various natural language generation tasks.
This work used a classifier to instruct the MCMC-based models where and how to refine the candidate sentences, and proposed a two-step approach, “Predict and Revise”, for constrained sentence generation.
This paper proposes inductive conformal prediction (ICP) algorithms for the tasks of text infilling and part-of-speech (POS) prediction for natural language data and designs new ICP-enhanced algorithms for POS tagging based on BERT and BiLSTM models.
Adding a benchmark result helps the community track progress.