A Systematic Evaluation of Large Language Models of Code
- 1. Carnegie Mellon University
Description
These are datasets for the paper:
"A Systematic Evaluation of Large Language Models of Code"
https://arxiv.org/pdf/2202.13169.pdf
The code is available at: https://github.com/VHellendoorn/Code-LMs
The file "unseen_test_sets.tar.gz" contains test sets of ~100 files in each of 12 programming languages.
These files are not included in The Pile, and thus models such as GPT-Neo, GPT-J, GPT-NeoX were not trained on them.
In the paper, we use these test sets to compare a variety of language models of code including OpenAI's Codex, GPT-J, GPT-Neo, GPT-NeoX-20B, and CodeParrot and our PolyCoder model.
The file "index.zip" includes an index of the training set file paths and commit SHAs.
The other files, such as "2-7B-150K.tar", are trained model checkpoints, as explained at https://github.com/VHellendoorn/Code-LMs .
Notes
Files
index.zip
Files
(16.3 GB)
Name | Size | Download all |
---|---|---|
md5:a4a367be4fbee217bca4e19ef203cf99
|
810.8 MB | Download |
md5:530400d496940107b1745a0317812795
|
810.8 MB | Download |
md5:ca615c1547288fe29b750992259dea8c
|
2.3 GB | Download |
md5:dbdba8f93e511cf9da3407d04c6c3066
|
5.6 GB | Download |
md5:9d503f66a809240b1d6d59b3b0569f40
|
5.6 GB | Download |
md5:eaf63ec6e69478bde42b1b28a46a1881
|
1.3 GB | Preview Download |
md5:6473b2cb1a475de40309832237922c2d
|
898.7 kB | Download |