3260 papers • 126 benchmarks • 313 datasets
Given unsupervised Language Modeling as pretraining task, the objective is to generate texts under particular control attributes (Topic, Sentiment)
(Image credit: Papersgraph)
These leaderboards are used to track progress in controllable-language-modelling-2
No benchmarks available.
Use these libraries to find controllable-language-modelling-2 models and implementations
No datasets available.
No subtasks available.
It is shown that automatic metrics provide a better guidance than human on discriminating system-level performance in Text Summarization and Controlled Generation tasks, and that multi-aspect human-aligned metric (UniEval) is not necessarily dominant over single-aspects human- aligned metrics (CTC, CtrlEval), and task-agnostic metrics (BLEU, BERTScore), particularly in Controlled generation tasks.
Adding a benchmark result helps the community track progress.