A system for easily mapping any natural language tasks into a human-readable prompted form and fine-tune a pretrained encoder-decoder model on this multitask mixture covering a wide variety of tasks.
Large language models have recently been shown to attain reasonable zero-shot generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that this is a consequence of implicit multitask learning in language models' pretraining (Radford et al., 2019). Can zero-shot generalization instead be directly induced by explicit multitask learning? To test this question at scale, we develop a system for easily mapping any natural language tasks into a human-readable prompted form. We convert a large set of supervised datasets, each with multiple prompts with diverse wording. These prompted datasets allow for benchmarking the ability of a model to perform completely held-out tasks. We fine-tune a pretrained encoder-decoder model (Raffel et al., 2020; Lester et al., 2021) on this multitask mixture covering a wide variety of tasks. The model attains strong zero-shot performance on several standard datasets, often outperforming models up to 16x its size. Further, our approach attains strong performance on a subset of tasks from the BIG-bench benchmark, outperforming models up to 6x its size. All trained models are available at https://github.com/bigscience-workshop/t-zero and all prompts are available at https://github.com/bigscience-workshop/promptsource.
Sheng Shen
9 papers
Urmish Thakker
4 papers
Zheng-Xin Yong
5 papers
Han Wang
4 papers
Leo Gao
4 papers
Stella Biderman
7 papers
Canwen Xu
6 papers
Colin Raffel
11 papers
Rachel Bawden
3 papers
Alexander M. Rush
18 papers
Zaid Alyafeai
5 papers
Andrea Santilli
3 papers
Matteo Manica
4 papers
Thomas Wang
4 papers
Lintang Sutawika
4 papers
Teven Le Scao
6 papers
M Saiful Bari
6 papers
Albert Webson
4 papers
Taewoon Kim
3 papers
Thomas Wolf
4 papers
Victor Sanh
4 papers
Stephen H. Bach
3 papers
Antoine Chaffin
2 papers
Arnaud Stiegler
2 papers
Arun Raja
2 papers
Manan Dey
2 papers
S. Sharma
2 papers
Eliza Szczechla
2 papers
Gunjan Chhablani
3 papers
Nihal V. Nayak
4 papers
Debajyoti Datta
2 papers
Jonathan D. Chang
1 papers
Mike Tian-Jian Jiang
2 papers
Harshit Pandey
3 papers
Trishala Neeraj
2 papers
Jos Rozen
2 papers
Abheesht Sharma
3 papers
Thibault Févry
5 papers
Jason Alan Fries
3 papers
R. Teehan
2 papers
T. Bers
2 papers