1
All the News That’s Fit to Fabricate: AI-Generated Text as a Tool of Media Misinformation
2
Language (Technology) is Power: A Critical Survey of “Bias” in NLP
3
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks
4
How Can We Accelerate Progress Towards Human-like Linguistic Generalization?
5
UnifiedQA: Crossing Format Boundaries With a Single QA System
6
Experience Grounds Language
7
StereoSet: Measuring stereotypical bias in pretrained language models
8
Adversarial Training for Large Neural Language Models
9
Pretrained Transformers Improve Out-of-Distribution Robustness
10
TTTTTackling WinoGrande Schemas
11
Train Large, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers
12
REALM: Retrieval-Augmented Language Model Pre-Training
13
How Much Knowledge Can You Pack into the Parameters of a Language Model?
14
Scaling Laws for Neural Language Models
15
Multilingual Denoising Pre-training for Neural Machine Translation
16
Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference
17
PIQA: Reasoning about Physical Commonsense in Natural Language
18
Reducing Sentiment Bias in Language Models via Counterfactual Evaluation
19
Automatic Detection of Generated Text is Easiest when Humans are Fooled
20
Adversarial NLI: A New Benchmark for Natural Language Understanding
21
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
22
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
23
NumNet: Machine Reading Comprehension with Numerical Reasoning
24
SummAE: Zero-Shot Abstractive Text Summarization using Length-Agnostic Auto-Encoders
25
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
26
A Constructive Prediction of the Generalization Error Across Scales
27
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
28
UNITER: UNiversal Image-TExt Representation Learning
29
UNITER: Learning UNiversal Image-TExt Representations
30
Technical report on Conversational Question Answering
31
TinyBERT: Distilling BERT for Natural Language Understanding
32
Fine-Tuning Language Models from Human Preferences
33
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
34
The Woman Worked as a Babysitter: On Biases in Language Generation
35
Release Strategies and the Social Impacts of Language Models
36
Natural Questions: A Benchmark for Question Answering Research
37
RoBERTa: A Robustly Optimized BERT Pretraining Approach
38
The CommitmentBank: Investigating projection in naturally occurring discourse
40
Probing Neural Network Comprehension of Natural Language Arguments
41
XLNet: Generalized Autoregressive Pretraining for Language Understanding
42
GLTR: Statistical Detection and Visualization of Generated Text
43
Reducing Gender Bias in Word-Level Language Models with a Gender-Equalizing Loss Function
44
Defending Against Neural Fake News
45
Fair Is Better than Sensational: Man Is to Doctor as Woman Is to Doctor
46
Story Ending Prediction by Transferable BERT
47
MASS: Masked Sequence to Sequence Pre-training for Language Generation
48
SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems
49
HellaSwag: Can a Machine Really Finish Your Sentence?
50
BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions
51
Unsupervised Data Augmentation for Consistency Training
52
Generating Long Sequences with Sparse Transformers
53
The Curious Case of Neural Text Degeneration
54
Improving Multi-Task Deep Neural Networks via Knowledge Distillation for Natural Language Understanding
55
Lipstick on a Pig: Debiasing Methods Cover up Systematic Gender Biases in Word Embeddings But do not Remove Them
56
DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs
57
Massively Multilingual Neural Machine Translation
58
Right for the Wrong Reasons: Diagnosing Syntactic Heuristics in Natural Language Inference
59
Multi-Task Deep Neural Networks for Natural Language Understanding
60
Learning and Evaluating General Linguistic Intelligence
61
Cross-lingual Language Model Pretraining
62
Transformer-XL: Attentive Language Models beyond a Fixed-Length Context
63
An Empirical Model of Large-Batch Training
64
Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks
65
ReCoRD: Bridging the Gap between Human and Machine Commonsense Reading Comprehension
66
Model Cards for Model Reporting
67
WiC: the Word-in-Context Dataset for Evaluating Context-Sensitive Meaning Representations
68
WiC: 10, 000 Example Pairs for Evaluating Context-Sensitive Representations
69
Dissecting Contextual Word Embeddings: Architecture and Representation
70
Meta-Learning for Low-Resource Neural Machine Translation
71
CoQA: A Conversational Question Answering Challenge
72
QuAC: Question Answering in Context
73
Can a Suit of Armor Conduct Electricity? A New Dataset for Open Book Question Answering
75
The Natural Language Decathlon: Multitask Learning as Question Answering
76
Know What You Don’t Know: Unanswerable Questions for SQuAD
77
A Simple Method for Commonsense Reasoning
78
Looking Beyond the Surface: A Challenge Set for Reading Comprehension over Multiple Sentences
79
Gender Bias in Coreference Resolution
80
A Call for Clarity in Reporting BLEU Scores
81
Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge
82
Annotation Artifacts in Natural Language Inference Data
83
Generating Wikipedia by Summarizing Long Sequences
84
Universal Language Model Fine-tuning for Text Classification
85
Deep Learning Scaling is Predictable, Empirically
86
Decoupled Weight Decay Regularization
87
Few-shot Autoregressive Density Estimation: Towards Learning to Learn Distributions
88
Learned in Translation: Contextualized Word Vectors
89
Attention is All you Need
90
TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension
91
RACE: Large-scale ReAding Comprehension Dataset From Examinations
92
Domain randomization for transferring deep neural networks from simulation to the real world
93
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
94
Learning to Optimize Neural Nets
95
Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer
96
Optimization as a Model for Few-Shot Learning
97
RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning
98
Sequence-Level Knowledge Distillation
99
The LAMBADA dataset: Word prediction requiring a broad discourse context
100
Learning to learn by gradient descent by gradient descent
101
Matching Networks for One Shot Learning
102
Adaptive Computation Time for Recurrent Neural Networks
103
Exploring the Limits of Language Modeling
104
Improving Neural Machine Translation Models with Monolingual Data
105
Semi-supervised Sequence Learning
106
Representation Learning Using Multi-Task Deep Neural Networks for Semantic Classification and Information Retrieval
107
Distilling the Knowledge in a Neural Network
108
A Neural Network for Factoid Question Answering over Paragraphs
109
GloVe: Global Vectors for Word Representation
110
Edinburgh’s Phrase-based Machine Translation Systems for WMT-14
111
Semantic Parsing on Freebase from Question-Answer Pairs
112
Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation
113
Efficient Estimation of Word Representations in Vector Space
114
Guide for Conducting Risk Assessments
115
The Winograd Schema Challenge
116
Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning
117
SentiWordNet 3.0: An Enhanced Lexical Resource for Sentiment Analysis and Opinion Mining
118
The Third PASCAL Recognizing Textual Entailment Challenge
120
Corpus-based Learning of Analogies and Semantic Relations
121
Combining Independent Modules to Solve Multiple-choice Synonym and Analogy Problems
122
Learning to Learn Using Gradient Descent
124
The Penn Treebank: Annotating Predicate Argument Structure
125
Information-Based Objective Functions for Active Data Selection
126
Acquiring a Single New Word
127
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
128
Multi-Agent Dual Learning
129
Language Models are Unsupervised Multitask Learners
130
An Adversarial Winograd Schema Challenge at Scale
131
Improving Language Understanding by Generative Pre-Training
132
Understanding backtranslation at scale
134
A corpus and evaluation framework for deeper understanding of commonsense stories
135
The Seventh PASCAL Recognizing Textual Entailment Challenge
136
The Sixth PASCAL Recognizing Textual Entailment Challenge
137
Natural language corpus data
138
NLTK: The Natural Language Toolkit
139
Machine Learning Challenges. Evaluating Predictive Uncertainty, Visual Object Classification, and Recognising Tectual Entailment
140
The Second PASCAL Recognising Textual Entailment Challenge
141
The PASCAL Recognising Textual Entailment Challenge
142
A Natural Logic Inference System
144
was an early advocate for scaling large generative likelihood models, and advised
145
Scholarship, Research, and Creative Work at Bryn Mawr College Scholarship, Research, and Creative Work at Bryn Mawr College