-
Notifications
You must be signed in to change notification settings - Fork 101
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
can not get the paper result #3
Comments
I hava the same question, what's your specific condition? When i train the model, the reward almost don't change. When i test the TMs, i find the training as if never learned sth. |
yes, learn nothing, but you should fix the TM when testing over the training stage |
Sorry to bother you, i don't really understand what u mean? How to fix the TMs? |
the author's code didn't do testing , so you have to write test code by yourself to get the fig result in paper. |
Sorry to bother you! |
Excuse me. Also have similar question, I can't understand why the state(TMs) and the new_state(TMs) are randomly generated in the step function which in Environment.py .It isn't meeting the logic of DRL. |
Can any get the result of the same as in paper as the model is not converging |
I have the same question. I dont't understand why the old state and the new state are randomly generated in the Environment.py. |
Did you run the whole simulations or not.
…On Wednesday, May 29, 2019, CZMG ***@***.***> wrote:
I have the same question. I dont't understand why the old state and the
new state are randomly generated in the Environment.py.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#3?email_source=notifications&email_token=AMC5WSNJIYYDIC37B2Y2SEDPXZ6D3A5CNFSM4GRGLWCKYY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODWPIXII#issuecomment-496929697>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AMC5WSOD2RZMH5N4WAZV5VDPXZ6D3ANCNFSM4GRGLWCA>
.
|
Excuse me. I run the whole simulations. But in my daily study, the STATE of Reinforcement Learning is usually changed by the ACTION, but in the code of this paper, we can find flie that in Environment.py, its NEW STATE and OLD STATE are randomly generated, which does not seem to meet the logic of Reinforcement Learning. Which teacher can answer my confusion? Thank you very much. |
I've also found this question. I think that the author need to do some explanations. It disobey the basic logic of reinforcement learning. @gissimo |
hello,Please ask how I can run the whole simulation, can you tell me the approximate steps, thank you very much! |
have you anyone get the fig2 result in paper? my model doesn't convergence。
The text was updated successfully, but these errors were encountered: