Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Not working? #1

Open
pmarcin92 opened this issue Dec 29, 2016 · 0 comments
Open

Not working? #1

pmarcin92 opened this issue Dec 29, 2016 · 0 comments

Comments

@pmarcin92
Copy link

Did anyone try to learn it so the agent really can play pong? I tried to learn it for over 30h on Tesla K80 and it doesn't look good at all.
I have also once concern about saving and restoring the learned weights. I modified the code to save the session once every 100000 iterations and I restore it like that:

    saver = tf.train.Saver()
    sess.run(tf.initialize_all_variables())
    saver = tf.train.import_meta_graph('pong-dqn-1300000.meta')
    saver.restore(sess, tf.train.latest_checkpoint('./'))

Is it me doing something wrong or there is a bug somewhere in the code preventing it from learn the pong?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant