1
Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
2
GenLine and GenForm: Two Tools for Interacting with Generative Language Models in a Code Editor
3
Evaluating Large Language Models Trained on Code
5
Implicit Representations of Meaning in Neural Language Models
6
Measuring Coding Challenge Competence With APPS
7
The Power of Scale for Parameter-Efficient Prompt Tuning
8
GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow
9
On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜
10
CodeXGLUE: A Machine Learning Benchmark Dataset for Code Understanding and Generation
11
Studying the Usage of Text-To-Text Transfer Transformer to Support Code-Related Tasks
12
Extracting Training Data from Large Language Models
13
Learning to Execute Programs with Instruction Pointer Attention Graph Neural Networks
14
PyMT5: Multi-mode Translation of Natural Language and Python Code with Transformers
15
Deep Just-In-Time Inconsistency Detection Between Comments and Source Code
16
BUSTLE: Bottom-up program-Synthesis Through Learning-guided Exploration
17
You Autocomplete Me: Poisoning Vulnerabilities in Neural Code Completion
18
Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data
19
Where should I comment my code? A dataset and model for predicting locations that need comments
20
DreamCoder: growing generalizable, interpretable knowledge with wake–sleep Bayesian program learning
21
Unsupervised Translation of Programming Languages
22
Language Models are Few-Shot Learners
23
Graph-based, Self-Supervised Program Repair from Diagnostic Feedback
24
IntelliCode compose: code generation using transformer
25
Hoppity: Learning Graph Transformations to Detect and Fix Bugs in Programs
26
Global Relational Models of Source Code
27
LambdaNet: Probabilistic Type Inference using Graph Neural Networks
28
OptTyper: Probabilistic Type Inference by Optimising Logical and Natural Constraints
29
Code Prediction by Feeding Trees to Transformers
30
Big Code != Big Vocabulary: Open-Vocabulary Models for Source Code
31
CodeBERT: A Pre-Trained Model for Programming and Natural Languages
32
Learning to Represent Programs with Property Signatures
33
Learning and Evaluating Contextual Embedding of Source Code
34
TypeWriter: neural type prediction with search-based validation
35
Learning to Fix Build Errors with Graph2Diff Neural Networks
36
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
37
CodeSearchNet Challenge: Evaluating the State of Semantic Code Search
38
SPoC: Search-based Pseudocode to Code
39
Write, Execute, Assess: Program Synthesis with a REPL
40
MathQA: Towards Interpretable Math Word Problem Solving with Operation-Based Formalisms
41
SequenceR: Sequence-to-Sequence Learning for End-to-End Program Repair
42
Deep learning type inference
43
Execution-Guided Neural Program Synthesis
44
Automatic Program Synthesis of Long Programs with a Learned Garbage Collector
45
Mapping Language to Code in Programmatic Context
46
SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing
47
NAPS: Natural Program Synthesis Dataset
48
DeepBugs: a learning approach to name-based bug detection
49
code2vec: learning distributed representations of code
50
Deep Contextualized Word Representations
51
Universal Language Model Fine-tuning for Text Classification
52
Learning to Represent Programs with Graphs
53
A Survey of Machine Learning for Big Code and Naturalness
55
Attention is All you Need
56
RobustFill: Neural Program Learning under Noisy I/O
57
Neural Sketch Learning for Conditional Program Generation
58
DeepCoder: Learning to Write Programs
59
Probabilistic model for code with decision trees
60
Hybrid computing using a neural network with dynamic external memory
61
Latent Predictor Networks for Code Generation
62
A Convolutional Attention Network for Extreme Summarization of Source Code
63
Exploring the Limits of Language Modeling
64
Automatic patch generation by learning correct code
65
Neural GPUs Learn Algorithms
66
Neural Random Access Machines
67
Semi-supervised Sequence Learning
68
Predicting Program Properties from "Big Code"
71
Phrase-Based Statistical Translation of Programming Languages
72
Towards a Big Data Curated Benchmark of Inter-project Code Clones
73
Code completion with statistical language models
74
Learning natural coding conventions
75
Structured Generative Models of Natural Source Code
76
Growing solver-aided languages with rosette
77
Syntax-guided synthesis
78
Lexical statistical machine translation for language migration
79
Mining source code repositories at massive scale using language modeling
80
On the naturalness of software
81
Generating Text with Recurrent Neural Networks
82
Automating string processing in spreadsheets using input-output examples
83
Combinatorial sketching for finite programs
84
On the synthesis of a reactive module
85
A Methodology for LISP Program Construction from Examples
86
Inferring LISP Programs From Examples
87
Knowledge and Reasoning in Program Synthesis
88
Toward automatic program synthesis
89
PROW: A Step Toward Automatic Program Writing
90
The FORTRAN automatic coding system
91
Prefix-Tuning: Optimizing Continuous Prompts for Generation
92
A large-scale benchmark for few-shot program induction and synthesis
93
Project CodeNet: A Large-Scale AI for Code Dataset for Learning a Diversity of Coding Tasks
94
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
95
Language Models are Unsupervised Multitask Learners
96
Learning Libraries of Subroutines for Neurally-Guided Bayesian Program Induction
97
Improving Language Understanding by Generative Pre-Training
98
A syntactic neural model for general-purpose code generation
99
GenProg: A Generic Method for Automatic Software Repair
100
Recurrent neural network based language model
101
Make sure the function signature is not unusual
102
We explore the semantic grounding of our models, investigating the extent to which these models can execute code given specific inputs (Section 6)
103
Alan Turing’s Electronic Brain: The Struggle to Build the ACE, the World’s Fastest Computer
104
when the model solves a task
105
limitations of our current model point toward interesting
106
’Consider the following Python function:\n\n{code}\n\n’ \ + ’This function solves the task: "{description}"\n\n’ \ + ’Fill in the ??? below:\n\n{tests}’