High training time for a model with QLSTM layer #896
-
Hello, I am trying to train a model in brevitas which comprises of an input QLSTM layer followed by Dense layers. The model has ~6K trainable parameters. I am facing issues with the training time of the model. In my case I am seeing training times in the range of ~5 hours per epoch. This seems to be quite high given the model configuration I use. As soon as I remove the QLSTM layer, the training times come back to some minutes per epoch. Is this issue common while training QLSTM layers? Is there a way to train models with this layer faster? Model Definition :
|
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 6 replies
-
You can try running training with the env flag BREVITAS_JIT=1 |
Beta Was this translation helpful? Give feedback.
I haven't had the chance to run your code, but on first glance, I don't see anything obviously wrong with your training script. I don't currently have the bandwidth to analyse your code any further.
This is a red flag to me that your network is not being compiled with PyTorch's JIT, with that is mind, I'd say your have 3 options: