This paper introduces CodeXGLUE, a benchmark dataset to foster machine learning research for program understanding and generation that includes a collection of 10 tasks across 14 datasets and a platform for model evaluation and comparison.
Benchmark datasets have a significant impact on accelerating research in programming language tasks. In this paper, we introduce CodeXGLUE, a benchmark dataset to foster machine learning research for program understanding and generation. CodeXGLUE includes a collection of 10 tasks across 14 datasets and a platform for model evaluation and comparison. CodeXGLUE also features three baseline systems, including the BERT-style, GPT-style, and Encoder-Decoder models, to make it easy for researchers to use the platform. The availability of such data and baselines can help the development and validation of new methods that can be applied to various program understanding and generation problems.
Junjie Huang
3 papers
Alexey Svyatkovskiy
1 papers
Ambrosio Blanco
2 papers
Colin B. Clement
3 papers
Dawn Drain
4 papers
Daxin Jiang
7 papers
Duyu Tang
5 papers
Ge Li
5 papers
Lidong Zhou
2 papers
Linjun Shou
7 papers
Long Zhou
5 papers
Michele Tufano
3 papers
Ming Gong
6 papers
Ming Zhou
14 papers
Nan Duan
17 papers
Neel Sundaresan
5 papers
Shao Kun Deng
1 papers
Shengyu Fu
3 papers
Shujie Liu
2 papers