2
Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
3
OPT: Open Pre-trained Transformer Language Models
4
PaLM: Scaling Language Modeling with Pathways
5
Do As I Can, Not As I Say: Grounding Language in Robotic Affordances
6
STaR: Bootstrapping Reasoning With Reasoning
7
Self-Consistency Improves Chain of Thought Reasoning in Language Models
8
Training language models to follow instructions with human feedback
9
Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?
10
Chain of Thought Prompting Elicits Reasoning in Large Language Models
11
Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model
12
LaMDA: Language Models for Dialog Applications
13
Scaling Language Models: Methods, Analysis & Insights from Training Gopher
14
Show Your Work: Scratchpads for Intermediate Computation with Language Models
15
Training Verifiers to Solve Math Word Problems
16
Multitask Prompted Training Enables Zero-Shot Task Generalization
17
Do Prompt-Based Models Really Understand the Meaning of Their Prompts?
18
Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing
19
Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity
20
GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow
21
Are NLP Models really able to Solve Simple Math Word Problems?
22
Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm
23
What Makes Good In-Context Examples for GPT-3?
24
Did Aristotle Use a Laptop? A Question Answering Benchmark with Implicit Reasoning Strategies
25
Making Pre-trained Language Models Better Few-shot Learners
26
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
27
Transformers: State-of-the-Art Natural Language Processing
28
It’s Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners
29
Language Models are Few-Shot Learners
30
Unsupervised Commonsense Question Answering with Self-Talk
31
5分で分かる!? 有名論文ナナメ読み:Jacob Devlin et al. : BERT : Pre-training of Deep Bidirectional Transformers for Language Understanding
32
PyTorch: An Imperative Style, High-Performance Deep Learning Library
33
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
34
Explain Yourself! Leveraging Language Models for Commonsense Reasoning
35
Attention is All you Need
36
Program Induction by Rationale Generation: Learning to Solve and Explain Algebraic Word Problems
37
Pointer Sentinel Mixture Models
38
Solving General Arithmetic Word Problems
39
MAWPS: A Math Word Problem Repository
40
Parsing Algebraic Word Problems into Equations
41
Learning to Solve Arithmetic Word Problems with Verb Categorization
42
A measure of intelligence
43
The structure of human intelligence: It is verbal, perceptual, and image rotation (VPR), not fluid and crystallized
44
GPT-J-6B: A 6 Billion Parameter Autoregressive Language Model. https://github.com/kingoflolz/mesh-transformer-jax
45
AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts
46
CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge
47
Language Models are Unsupervised Multitask Learners
48
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
49
The Cattell-Horn-Carroll Theory of Cognitive Abilities: Past, Present, and Future.
50
Heuristics and Biases: Individual Differences in Reasoning: Implications for the Rationality Debate?
51
Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)?
52
(a) Did you state the full set of assumptions of all theoretical results
53
with respect to the random seed after running experiments multiple times)? [No] Our paper mainly used GPT-3 API with greedy decoding
54
Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation
55
Do the main claims made in the abstract and introduction accurately reflect the paper's contributions and scope?
56
Have you read the ethics review guidelines and ensured that your paper conforms to them
57
code, data, models) or curating/releasing new assets... (a) If your work uses existing assets
58
Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable?
59
If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots
60
Did you discuss any potential negative societal impacts of your work