Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Gymnasium support #192

Draft
wants to merge 30 commits into
base: master
Choose a base branch
from
Draft
Changes from 1 commit
Commits
Show all changes
30 commits
Select commit Hold shift + click to select a range
a413077
Adds gymnasium to the setup
prabhatnagarajan Apr 15, 2023
35a420c
gym -> gymnasium
prabhatnagarajan Apr 15, 2023
e1d7ead
modifies calls to env step to use truncations
prabhatnagarajan Apr 15, 2023
c804fe3
some Atari changes
prabhatnagarajan Apr 15, 2023
5daca4c
Makes more env modifications
prabhatnagarajan Apr 15, 2023
85fe46f
Fixes some observations, and uses new Gym AtariEnv properly
prabhatnagarajan Apr 16, 2023
c7d62f7
makes some evaluator updates
prabhatnagarajan Apr 16, 2023
b51ae32
Gets evaluations working by modifying RandomizeAction class
prabhatnagarajan Apr 16, 2023
ffdc311
fixes setup
prabhatnagarajan Apr 17, 2023
07c464b
Adds a generic GymWrapper
prabhatnagarajan Apr 18, 2023
675c978
Shifts Pendulum version in example to v1 since v0 is deprecated
prabhatnagarajan Apr 23, 2023
85c38e1
Adds Q value computation to DDQN (and by extension DDQN)
prabhatnagarajan Apr 27, 2023
02048ae
removes filelock from setup
prabhatnagarajan May 10, 2023
98a4efc
removes all required items
prabhatnagarajan May 10, 2023
4818545
fixes setup
prabhatnagarajan Jun 24, 2023
e3c867c
merges with master
prabhatnagarajan Dec 28, 2023
3f98ef7
does gymnasium all to gymnasium atari
prabhatnagarajan Dec 28, 2023
b145609
Fixes multiprocessvector_env step
prabhatnagarajan Apr 2, 2024
0c770b5
Multiprocess fixes
brett-daley Apr 2, 2024
e82663e
Merge pull request #6 from brett-daley/gymnasium_support
prabhatnagarajan Apr 2, 2024
04c1dd5
OpenAI -> Farama Foundation
prabhatnagarajan Apr 3, 2024
c408b08
Makes modifications for gymnasium imports, etc.
prabhatnagarajan Apr 3, 2024
03f203f
Removes continuing_time_limit now that gymnasium has truncation
prabhatnagarajan Apr 3, 2024
5614323
Remove Monitor
prabhatnagarajan Apr 3, 2024
4b0494e
Removes things from __init__
prabhatnagarajan Apr 6, 2024
e92cc63
Moves gym folder in examples to gymnasium
prabhatnagarajan Apr 6, 2024
420dddb
Fixes some imports and some tests
prabhatnagarajan Apr 10, 2024
74198b9
Fixes Randomize Action Wrapper
prabhatnagarajan May 7, 2024
3237561
merges with main
prabhatnagarajan Jul 26, 2024
6f0eac6
Merge branch 'master' into gymnasium_support
prabhatnagarajan Aug 4, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Fixes some imports and some tests
  • Loading branch information
prabhatnagarajan committed Apr 10, 2024
commit 420dddb22e7036439c3ea21da7cf450b7adf9d8b
2 changes: 1 addition & 1 deletion pfrl/envs/abc.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
import numpy as np
import gymnasium as gym
from gym import spaces
from gymnasium import spaces

from pfrl import env

24 changes: 12 additions & 12 deletions tests/experiments_tests/test_evaluator.py
Original file line number Diff line number Diff line change
@@ -21,7 +21,7 @@ def test_evaluator_evaluate_if_necessary(save_best_so_far_agent, n_steps, n_epis
agent.get_statistics.return_value = []

env = mock.Mock()
env.reset.return_value = "obs"
env.reset.return_value = "obs", {}
env.step.return_value = ("obs", 0, True, False, {})
env.get_statistics.return_value = []

@@ -110,7 +110,7 @@ def test_async_evaluator_evaluate_if_necessary(save_best_so_far_agent, n_episode
agent.get_statistics.return_value = []

env = mock.Mock()
env.reset.return_value = "obs"
env.reset.return_value = "obs", {}
env.step.return_value = ("obs", 0, True, False, {})
env.get_statistics.return_value = []

@@ -179,10 +179,10 @@ def test_run_evaluation_episodes_with_n_steps(n_episodes, n_steps):
# Second episode: 4 -> 5 -> 6 -> 7 (done)
env.reset.side_effect = [("state", 0), ("state", 4)]
env.step.side_effect = [
(("state", 1), 0.1, False, {}),
(("state", 2), 0.2, False, {}),
(("state", 3), 0.3, False, {"needs_reset": True}),
(("state", 5), -0.5, False, {}),
(("state", 1), 0.1, False, False, {}),
(("state", 2), 0.2, False, False, {}),
(("state", 3), 0.3, False, True, {"needs_reset": True}),
(("state", 5), -0.5, False, False, {}),
(("state", 6), 0, False, False, {}),
(("state", 7), 1, True, False, {}),
]
@@ -226,12 +226,12 @@ def test_needs_reset(self):
# Second episode: 4 -> 5 -> 6 -> 7 (done)
env.reset.side_effect = [("state", 0), ("state", 4)]
env.step.side_effect = [
(("state", 1), 0, False, {}),
(("state", 2), 0, False, {}),
(("state", 3), 0, False, {"needs_reset": True}),
(("state", 5), -0.5, False, {}),
(("state", 6), 0, False, {}),
(("state", 7), 1, True, {}),
(("state", 1), 0, False, False, {}),
(("state", 2), 0, False, False, {}),
(("state", 3), 0, False, True, {"needs_reset": True}),
(("state", 5), -0.5, False, False, {}),
(("state", 6), 0, False, False, {}),
(("state", 7), 1, True, False, {}),
]
scores, lengths = evaluator.run_evaluation_episodes(
env, agent, n_steps=None, n_episodes=2