-
# Deep Q-Network (DQN) on LunarLander-v2 | Chan`s Jupyter
In this post, We will take a hands-on-lab of Simple Deep Q-Network (DQN) on openAI LunarLander-v2 environment. This is the coding exercise fr…
-
I tried to benchmark the follwing environments ['BipedalWalker-v2', 'BipedalWalkerHardcore-v2', 'CarRacing-v0', 'LunarLander-v2', 'LunarLanderContinuous-v2'] using ['A3C', 'DDPG', 'TD3', 'SAC', 'PG',…
-
I followed all the instructions here: https://spinningup.openai.com/en/latest/user/installation.html#check-your-install , but I have a problem with mpi4py. I tried the suggestion in stackoverflow (htt…
-
Describe the Question
Hello,
I have two questions after reading your paper.First, you used two Q-functions QA & QB .Every training iteration, QA and QB are swapped,Is this like Delayed network that…
-
I followed the install steps, but when I try to check my install with this following line, I get an "illegal hardware instruction" error.
`python -m spinup.run ppo --hid "[32,32]" --env LunarLande…
-
I am a beginner in RL and running env.render() doesn't open any environment window, please help.
environment_name = "CartPole-v1"
env = gym.make(environment_name)
episodes = 5
for episode in r…
-
Hello, first of all thanks for sharing your amazing job. I want to know Why I can't converge the network when I use your default setting(terminal: python run_atari_dqn.py)? the average 100 reward valu…
-
Hi,
I'm trying to create my own agent, following the example here: https://github.com/aunum/gold/tree/master/pkg/v1/agent/reinforce It's an example for `CartPole-v0`, while I'm playing around with …
-
configs should now be dictionaries within the overarching `Configs` dictionary, like so:
```python
Configs = {
'config_name1': {
'env': 'Acrobot-v1',
'gamma': 0.99
…
-
_**Is_ your feature request related to a problem? Please describe.**
I really appreciate your coding because it helped me a lot.
I am just a beginner of RL and I wonder if I can use ppo_rnd_tenso…