pip install -e .
python run_PointEnv.py configs/config_PointEnv.py
policy: no search
start: [0.03271197 0.99020872]
goal: [0.81310241 0.028764 ]
steps: 300
----------
policy: search
start: [0.03271197 0.99020872]
goal: [0.81310241 0.028764 ]
steps: 127
policy: no search
start: [0.03271197 0.99020872]
goal: [0.81310241 0.028764 ]
steps: 300
----------
policy: search
start: [0.03271197 0.99020872]
goal: [0.81310241 0.028764 ]
steps: 111
Initial SparseSearchPolicy (|V|=202, |E|=1894) has success rate 0.20, evaluated in 14.26 seconds
Filtered SparseSearchPolicy (|V|=202, |E|=986) has success rate 0.80, evaluated in 8.44 seconds
Took 10000 cleanup steps in 84.45 seconds
Cleaned SparseSearchPolicy (|V|=202, |E|=955) has success rate 1.00, evaluated in 6.69 seconds
- https://github.com/scottemmons/sgm
- https://github.com/google-research/google-research/tree/master/sorb
- https://github.com/sfujim/TD3
[1]: Michael Laskin, Scott Emmons, Ajay Jain, Thanard Kurutach, Pieter Abbeel, Deepak Pathak, "Sparse Graphical Memory for Robust Planning", 2020.
[2]: Benjamin Eysenbach, Ruslan Salakhutdinov, Sergey Levine, "Search on the Replay Buffer: Bridging Planning and Reinforcement Learning", 2019.