Skip to content

A repo with some utilities for evaluating LMs on code problems.

License

Notifications You must be signed in to change notification settings

laetitia-teo/code-eval

Repository files navigation

Evaluation scripts for code models

Compute quality

python experiments/compute_quality --config-name {conf}

Create random splits and finetune models on them

python experiments/dataset_random_split.py

Add the -r (--run) flag to run the experiment immediately (on a cluster).

Create quality-based splits and finetune models on them

python experiments/dataset_quality_split.py

Args:

-d --data-path path to the desired dataset
-g --gpu preferred gpu (one of {a100, v100})
-q --quality-key what quality key to use for the split
-r --run whether to run the experiment with slurm immediately after the split

Use the -q flag to change the key of the quality metric we use to split (we assume higher is better).

Train a model

python run.py --config-name {conf}

About

A repo with some utilities for evaluating LMs on code problems.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published