-
Notifications
You must be signed in to change notification settings - Fork 325
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature Request] Simplify the example and tutorial codes #1843
Comments
Hi @whatdhack Let me bring a few datapoints to move the discussion forward:
I hope that helps. If you (or anyone else) have concrete suggestions of what would be a clear and concise tutorial to add to the lib we'd be excited to get started working on it! |
I'm closing this per lack of feedback but if there's any actionable we can do I'll be thrilled to consider it! |
Specially in the dqn example, some of the well established logical division of Deep Learning are not followed. Hard to rationalize why the SyncDataCollector has a policy network attached to it. Also, looks like the dqn example calls the MLP 3 times in one iteration. !! |
It's very much WIP but here's the PR that will hopefully clarify things There's a link on top to see the doc rendered according to this work |
Motivation
It is hard to follow and understand the example and tutorials. As an example, if I compare the 2 flavors of cartpole PyTorch code, the one from PyTorch pytorch/tutorial is far easier to understand and follow than the one in pytorch/rl.
https://pytorch.org/tutorials/intermediate/reinforcement_q_learning.html
https://github.com/pytorch/rl/blob/main/examples/dqn/dqn_cartpole.py
Solution
A clear and concise example code.
Alternatives
Additional context
Checklist
The text was updated successfully, but these errors were encountered: