Error running eval.py
on checkpoint created using --pretrained_lm=gpt2
.
#71
Labels
bug
Something isn't working
Here's the steps to recreate:
First, to show that it's specifically related to the
--pretrained-lm
argument, run this train/eval pair once without--pretrained-lm=gpt2
in the training arguments, then run the pair again with--pretrained-lm=gpt2
in the training arguments.When you run the above with the
--pretrained_lm=gpt2
argument, you get the following error message:Doing some archeology, we find at one point in time
pretrained_lm
was removed from the training args before evaluation.That change was then modified to only delete the
--pretrained_lm
argument if--lora
was passed.The logic explicitly raises if
lora
_, and notpretrained_lm
.The logic implicitly fails if
pretrained_lm
and notlora
.I'm not sure if that's purposeful or accidental. I don't know much about lora and how it interacts with the model. I'm guessing it's fine to run
pretrained_lm
withoutlora
.Just logging this right now as research/investigation notes to pick back up later.
The text was updated successfully, but these errors were encountered: