-
Hi Philip, I was wondering whether it's possible to manually set the emulator speed. It'd be nice to further increase the speed, say to 5000%, during training. Additionally, when demoing the RL agent,…
-
Hi Denny
Again, I do appreciate your work!
I was thinking of implementing DQN with **Dyna-Q** Algorithm where the **Q(s,a)** is updated not only by **real** experience, but also by **simulated** ex…
-
Hi,
why did you include the minus sign in the `grad_ys` argument of the bottom function?
`self.parameters_gradients = tf.gradients(self.action_output,self.parameters,-self.q_gradient_input/BATCH_SIZ…
-
Setting up openai/universe, I used the "universe starter agent" as a smoke test.
After adjusting the number of workers to better utilize my CPU, I saw the default PongDeterministic-v3 start winnin…
-
Could you pinpoint the code where actor's parameters (weights) are being updated ?
I am particularly looking for the step where gradient of critic is calculated wrt to action variables and actor wrt …
-
Greetings!
We happen to have just pushed into the open source one of the Lasagne-based library for reinforcement learning algorithm design.
- The repo's here: [AgentNet](https://github.com/yandexdatas…
-
## Configuration:
- Untuned: 10 repeats, and median stability score is taken. Default parameters.
- Tuning: On Spark, 300 evaluations
## Results:
- Delta, Positive, tuning is better.
- Different Combi…
-
Great project! I'm looking to use this with a Kinect v2 camera for a robotics application. I have 26 different joints each with x,y,z coordinates that will be my state space. Looking through the code …
-
Please find below the feedback from your work done in the 1st iteration.
Wiki
Homepage (Vision)
- Limitations can be set in order to define specifically the purpose of the application, e.g., it can b…