-
Even after a bigger run, agents don't learn:
according to the pressurplate we have a reward in [-0.9,0] if the agent is in the same room of the assigned plate and reward [-1,...,-N] otherwise.
I tri…
-
Hi Danijar,
I am currently trying to use higher image resolutions like 256x256 for Dreamer. By simply changing the resolution e.g. for DM control suite, JAX is not able to trace/compile the trainin…
-
Dear Simar et al.,
First of all, I would like to thank you for your research. I believe it is very well done and deserves to be studied carefully to learn from your perspectives, methods, and insig…
-
Hi! I'm trying to implement DDPG as well based on paper [Continuous control with deep reinforcement learning](http://arxiv.org/pdf/1509.02971.pdf). Though without much success yet... So I was looking …
-
Add support capabilities for the environment to accept multi agent.
-
## DDPG training logs
-
**Describe the bug**
Describe the bug
In DeepSpeed-Chat step3, a runtime error: The size of tensor a (4) must match the size of tensor b (8) at non-singleton dimension 0 will be thrown when inferenc…
-
Hi,
I have used MultiInputProcessor with DQN and it works fine. I am trying to use that feature to train an agent using DDPG.
I have 3 inputs from my environment: Image, two 1D vectors of size (1,3)…
-
https://tominute.github.io/2018/10/19/%E5%B0%8F%E7%99%BD%E7%AC%94%E8%AE%B0-Real-time-Actor-Critic-Tracking/
研究僧写字的地方
-
Dear author, I am very interested in your work, may I ask how long you run an experiment?