You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, Thank you very much for sharing your dataset. We tried to reproduce the BC training on lamp_low and only achieved ~3% success rate. I used the following arguments:
Unfortunately, we and many others find that imitation learning training is not very predictable or monotonous. Most of the time, one has to save and evaluate a series of checkpoints throughout training to find the best one—this varies from run to run.
That said, I don't see any obvious differences between your HPs there and the ones I used, except for the batch size of 128, where I typically used 256 (or higher) for state-based training.
So, I'd suggest you test slightly larger batch sizes and a couple more seeds and evaluate more checkpoints throughout training. I'm also in the process of releasing the weights for the models I trained.
Hi, Thank you very much for sharing your dataset. We tried to reproduce the BC training on lamp_low and only achieved ~3% success rate. I used the following arguments:
what parameters should we adjust to achieve 7% success rate as listed in paper, or it's just a margin of error that we are experiencing?
The text was updated successfully, but these errors were encountered: