1
Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
2
Can language models learn from explanations in context?
3
One-Shot Learning from a Demonstration with Hierarchical Latent Language
4
Training language models to follow instructions with human feedback
5
PromptSource: An Integrated Development Environment and Repository for Natural Language Prompts
6
ZeroPrompt: Scaling Prompt-Based Pretraining to 1, 000 Tasks Improves Zero-Shot Generalization
7
UnifiedSKG: Unifying and Multi-Tasking Structured Knowledge Grounding with Text-to-Text Language Models
8
ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning
9
MetaICL: Learning to Learn In Context
10
Multitask Prompted Training Enables Zero-Shot Task Generalization
11
FILM: Following Instructions in Language with Modular Methods
12
Reframing Instructional Prompts to GPTk’s Language
13
Finetuned Language Models Are Zero-Shot Learners
14
FLEX: Unifying Evaluation for Few-Shot NLP
15
Disfl-QA: A Benchmark Dataset for Understanding Disfluencies in Question Answering
16
Cross-Task Generalization via Natural Language Crowdsourcing Instructions
17
CrossFit: A Few-shot Learning Challenge for Cross-task Generalization in NLP
18
The Power of Scale for Parameter-Efficient Prompt Tuning
19
Adapting Language Models for Zero-shot Learning by Meta-tuning on Dataset and Prompt Collections
20
CaSiNo: A Corpus of Campsite Negotiation Dialogues for Automatic Negotiation Systems
21
Learning to Generate Task-Specific Adapters from Task Description
22
Author’s Sentiment Prediction
23
Learning from Task Descriptions
24
mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer
25
Language-Conditioned Imitation Learning for Robot Manipulation Tasks
26
The Turking Test: Can Language Models Understand Instructions?
27
Transformers: State-of-the-Art Natural Language Processing
28
Understanding Points of Correspondence between Sentences for Abstractive Summarization
29
Natural language to SQL: Where are we today?
30
Language Models are Few-Shot Learners
31
UnifiedQA: Crossing Format Boundaries With a Single QA System
32
Fast Domain Adaptation for Goal-Oriented Dialogue Using a Hybrid Generative-Retrieval Transformer
33
Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension
34
ALFRED: A Benchmark for Interpreting Grounded Instructions for Everyday Tasks
35
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
36
Meta-Dataset: A Dataset of Datasets for Learning to Learn from Few Examples
37
The Natural Language Decathlon: Multitask Learning as Question Answering
38
Deep Learning Scaling is Predictable, Empirically
39
Harvesting Common-sense Navigational Knowledge for Robotics from Uncurated Text Corpora
40
Revisiting Unreasonable Effectiveness of Data in Deep Learning Era
41
The E2E Dataset: New Challenges For End-to-End Generation
42
JFLEG: A Fluency Corpus and Benchmark for Grammatical Error Correction
43
A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks
44
The Language Demographics of Amazon Mechanical Turk
45
Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning
46
A unified architecture for natural language processing: deep neural networks with multitask learning
47
ROUGE: A Package for Automatic Evaluation of Summaries
48
Scaling to Very Very Large Corpora for Natural Language Disambiguation
50
OHSUMED: an interactive retrieval evaluation and new large test collection for research
51
Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
52
An Adversarial Winograd Schema Challenge at Scale
53
The Sixth PASCAL Recognizing Textual Entailment Challenge
54
The PASCAL Recognising Textual Entailment Challenge