Hi @mallamanis ,
How are you?
Thanks for sharing the code!
I'm trying to reproduce the PLDI dataset +results.
I am preparing the data according to: https://github.com/typilus/typilus/blob/master/src/data_preparation/README.md.
I ran bash scripts/prepare_data.sh
and modified it to have the line
bash /usr/src/datasetbuilder/scripts/clone_from_spec.sh /usr/src/datasetbuilder/pldi2020-dataset.spec
instead of the two loops that start with: while IFS= read -r line
and for repo in ./*; do
.
This is running for 5 days already, and pytype-single
is the process that currently runs.
The last three lines that were printed to console are:
[119/411] infer google.auth.jwt
[120/411] infer google.oauth2._client
[121/411] infer google.oauth2.service_account
Is this expected?
Does that mean that about 30% of the work is done, and it will take approximately 10-12 more days?
This is running in an Ubuntu machine with many cores and enough RAM (top
shows that only a single core is utilized, and 37% of the RAM).
Thanks!