You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Sep 24, 2024. It is now read-only.
The quantization parameter bnb_4bit_compute_dtype seem to be optional for fine-tuning, but when we run lm-harness we get an error if it is not specified in the config.
This is a good opportunity to use Pydantic @model_validator to validate the combination of parameters on the configuration, ensuring a fail-early approach to invalid choices.
That is good to know!
I got the error in lm-harness eval though. What puzzled me was that I got no errors during fine-tuning but I got one during eval - perhaps that's something required at inference time while it is somehow implied by HF code at tuning time?
Interesting, I'm not sure. Is it possible that you were already passing the LoftQ config during finetuning and that was handling quantization for you? I'm not positive about the defaults used by HuggingFace when not explicitly specified.
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
The quantization parameter
bnb_4bit_compute_dtype
seem to be optional for fine-tuning, but when we run lm-harness we get an error if it is not specified in the config.Example: http://10.145.91.219:8265/#/jobs/raysubmit_RvS6DHdgeMPYQjW9
Solution: the code runs properly if we add the parameter to the config, e.g.
Shall we always put a default value (e.g.
bfloat16
) or just document it to make sure it is specified when needed?The text was updated successfully, but these errors were encountered: