-
Notifications
You must be signed in to change notification settings - Fork 72
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[QUESTION] Why does training speed go down? #158
Comments
Hi zouharvi, I noticed this behavior as well. I think it has something to do with "Encoder model fine-tuning". After this the speed gradually decreases for me from 13.98it/s to 5.85it/s at the end of the epoch. Could someone comment if this is an excepted behavior? |
Indeed, without encoder fine-tuning ( |
hmmm and what happens on the second epoch? I actually never noticed this... |
In the second and the next epochs it converges to ~5it/s for me (A10G with batch size 6). |
Hi, I trained two reference-free QE models on in-domain data with 300k segments. One with |
I noticed that
comet-train
(after encoder finetuning) has speed of ~12it/s at e.g. 30% which drops to ~7it/s at 60% and to ~6it/s at 90% of the epoch.I'm using NVIDIA A10G GPUs and the following software versions:
The text was updated successfully, but these errors were encountered: