-
Notifications
You must be signed in to change notification settings - Fork 346
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reimplementation in RL4CO #58
Comments
Hi! Thanks for bringing it to my attention, I will definitely check it out! Are you able to reproduce the results from the paper with your implementation (training time, evaluation time and performance?). |
Thanks for your quick answer 🚀
We would be more than happy to address your feedback if you check out RL4CO, you may contact us any time 😄 |
(Late) edit: now we are way more efficient as explained here! |
Great! I have added a link in the readme. However, I wonder if you have also had a look at https://github.com/cpwan/RLOR, they claim a 8x speedup over this repo using PPO. |
Yes, we are aware of it! |
Hi there 👋🏼
First of all, thanks a lot for your library, it has inspired several works in our research group!
We are actively developing RL4CO, a library for all things Reinforcement Learning for Combinatorial Optimization. We started the library by modularizing the Attention Model, which is the basis for several other autoregressive models. We also used some recent software (such as TorchRL, TensorDict, PyTorch Lightning and Hydra) as well as routines such as FlashAttention, and made everything as easy to use as possible in the hope of helping practitioners and researchers.
We welcome you to check RL4CO out ^^
The text was updated successfully, but these errors were encountered: