This repository is a toolkit to do machine learning for programming languages. It implements tokenization, dataset preprocessing, model training and model evaluation.
We provide reference implementations of the following papers:
- DOBF: A Deobfuscation Pre-Training Objective for Programming Languages (2021)
- TransCoder: Unsupervised Translation of Programming Languages (2020)
We also provide pre-trained models for language modeling, translation and deobfuscation.
Dependencies
Run install_env.sh. We use black code formatter.
Data
Source code processors
This repository contains programming languages processors for C++, Java and Python. These processors include:
- tokenization and detokenization
- obfuscation
- function extractions
These processors are based on TreeSitter parsers. As these parsers are available in more than 30 programming languages, one can easily create a new programming language processor.
Example of code tokenization:
from codegen_sources.preprocessing.lang_processors.java_processor import JavaProcessor
java_code = r"""class HelloWorld {
public static void main(String[] args) {
System.out.println("Hello, World!");
}
}"""
java_processor = JavaProcessor(root_folder="<YOUR_TREESITER_FOLDER>")
tokenized_java_code = java_processor.tokenize_code(java_code)
print(tokenized_java_code)
BPE
This repository provides wrappers for fast BPE and Roberta BPE at file level.
Dataset Preprocessing
This repository contains a pipeline to create programming languages datasets. Now it supports four datasets modes:
- Monolingual (ex: Java source code)
- Monolingual Functions (ex: Java functions)
- Monolingual Obfuscated (ex: Obfuscated Java source code. [Details here])
- Monolingual Obfuscated Functions (ex: Obfuscated Java functions)
First, download C++ / Java / Python source code from Google BigQuery. To run our preprocessing pipeline, you need to donwload the raw source code on your machine in a JSON format. A sample of it is given here.
The pipeline does the following:
- Source code extraction from json (
.json.gz
) and tokenization (.tok
) - Train BPE codes and vocab
- Apply BPE (
.bpe
) - Binarization (
.pth
) - Symlink folder with appropriate file names for
.pth
(XLM-syml). To be given asdata_path
argument for training.
To run the pipeline :
python -m codegen_sources.preprocessing.preprocess \
<DATA_PATH> \ # folder containing json.gz
--langs java cpp python \ # languages to process
--mode monolingual_functions \ # dataset mode
--bpe_mode=fast_bpe \ # BPE mode. by default it is fast_BPE. can be roberta_bpe
--local=True \ # Run on your local machine if True. If False run on a cluster (requires submitit setup)
--train_splits=1 # Number of trainings splits
If you give several languages, the BPE codes and vocab will be learned commonly on these languages , so that you will have a common vocabulary to train one model for several languages. If you do not want that, launch the pipeline on every language separatly. These tests test the pipeline on different modes. It will give you an overview of the possible options.
Also, we provide the BPE codes and vocabulary here. These are the codes and vocabulary used for TransCoder and DOBF. They were learned on concatenated C++, Java, and Python data. If you want to use them instead of learning new ones, give the corresponding paths as fastbpe_code_path
and fastbpe_vocab_path
arguments.
In TransCoder and DOBF readmes, we provide the commands to preprocess the respective datasets.
Model
Overview
In this repository, we provide code to train transformer-based models (code based on XLM repository). The available training tasks are the following:
- Masked Language Model (MLM)
- Causal Language Model (CLM)
- Supervised Machine translation (MT)
- Classification
- Deobfuscation = DOBF
- Unsupervised Machine translation = TransCoder (Denoising auto encoding AE + Back Translation BT)
We evaluate our models with metrics adapted to each task (e.g. computation accuracy and BLEU score for TransCoder, subtoken score for Deobfuscation).
Also, we provide wrappers to fine-tune and evaluate our models on CodeXGLUE benchmark.
Download models
You can donwload the following models :
- MLM
- TransCoder. Use it to translate some code here.
- DOBF. Use it to deobfuscate some code here.
Re train specific models
To have details on how to retrain specific models, please refer to the README specific to each model.
References
TransCoder model (NeurIPS 2020)
[1] B. Roziere*, M.A. Lachaux*, L. Chanussot, G. Lample Unsupervised Translation of Programming Languages.
@article{roziere2020unsupervised,
title={Unsupervised translation of programming languages},
author={Roziere, Baptiste and Lachaux, Marie-Anne and Chanussot, Lowik and Lample, Guillaume},
journal={Advances in Neural Information Processing Systems},
volume={33},
year={2020}
}
DOBF
[2] B. Roziere*, M.A. Lachaux*, M. Szafraniec , G. Lample DOBF: A Deobfuscation Pre-Training Objective for Programming Languages.
@article{roziere2021dobf,
title={DOBF: A Deobfuscation Pre-Training Objective for Programming Languages},
author={Roziere, Baptiste and Lachaux, Marie-Anne and Szafraniec, Marc and Lample, Guillaume},
journal={arXiv preprint arXiv:2102.07492},
year={2021}
}
* Equal Contribution
License
CodeGen
is under the license detailed in the Creative Commons Attribution-NonCommercial 4.0 International license. See LICENSE for more details.