-
Dear author; thank you for your great work.
I have 2 question about sb3_highway_dqn example, please guide me:
1- How can I extract graphs related to reinforcement learning? (I mean the 3 presented…
-
@eleurent Hi Edouard,
I've trained my model in both Highway and Intersection environment. All the hyperparameters are as same as each other. And I used DQN (MlpPolicy) for both of them. But the pro…
-
Given a traffic network with multiple intersections, I want to deploy a single DQN to all the individual traffic light signal. How can I do that?
-
- [ ] Hyperparameter tweaking
- [x] Addition of "target value network"
- [ ] Replay priority
- [ ] Move to Dueling Deep Q
- Research other improvements/extensions that have been made to DQN netw…
-
It would be really nice if we could have an example that re-implements the original [DQN ](https://www.cs.toronto.edu/~vmnih/docs/dqn.pdf) paper exactly. The Ms. PacMan one is pretty close but AFAIK i…
-
- consider implementing with openai's battlesnake env that amazon created. A tutorial for using the environment can be found in the following links:
- [awslabs-battlesnake-ai repo](https://github.c…
-
Train the initial DQN model. This model will have 2 actions: close or hold. Hyperparameter search for different numbers of memories, evaluate convergence
-
### What happened + What you expected to happen
I am trying to write a custom policy with a postprocess_trajectory to post-process infos. However, after one training iteration, the infos in the raw…
-
-
I noticed that there are two model configuration parameters: fcnet_hiddens in model_config and hiddens in the configuration of dqn.py.
I try to print the whole model (model.base_model.summary()), it…