-
Hi,
I'm trying to run your code in DVRL mode (according to the configurations you mention in the README file) and the results are significantly lower compared to the ones published in your paper (a…
-
### 🐛 Bug
Hello,
I wrote a customized DQN policy trying to use Large Language Model to modify q-value before the dqn policy predicts an action. The idea is simple: every time the agent gets an obs…
-
First of all, thanks for the amazing implementation, really helpful for understanding the DQN.
I'm curious about the results on the space invaders, it shows the avg 2772, while PER original paper sho…
-
I'm working on the DQN code but overall it's still 10x slower than torch. I've narrowed the issue down to a line with only an `argmax`. Can anyone help me fix this?
On an m2 MAX 96gb.
Here's th…
-
I wanted to know if a contribution is welcomed here, and if it is, how to contribute? I mean, is there any guideline for how we should implement agents?
In fact, I wanted to implement agents like D…
-
Hi!
Let's bring the reinforcement learning course to all the Korean-speaking community 🌏 (currently 9 out of 77 complete)
Would you want to translate? Please follow the 🤗 [TRANSLATING guide](ht…
-
Hi!
Let's bring the reinforcement learning course to all the Russian-speaking community 🌏
Would you want to translate? Please follow the 🤗 [TRANSLATING guide](https://github.com/huggingface/tran…
-
*please fill this in*
-
## Motivation
I'm trying to create an RL agent which works with graph data structures. but when I pass the torch_geomteric `Data` object to `TensorDict`, It says
```
we only supports tensorcla…
-
I use tx1 and follow the 8 steps, but failed. 16G is enough for this project?
First time , I get the note like this
You are using pip version 9.0.1, however version 10.0.1 is available.
You shou…
yi2ha updated
6 years ago