-
Hi,
for the last couple of days I have been working on scripts for training and evaluating RL agents for multiplayer dominoes, but I have faced two main issues:
1. **Scale**: Any training seems …
-
Hi Denny,
Thanks for this wonderful resource. It's been hugely helpful. Can you say what your results are when training the DQN solution? I've been unable to reproduce the results of the DeepMind p…
-
## 🐛 Bug
The memory usage comparison between the same data structures implemented with different backends (PyTorch tensors and NumPy arrays) shows over 4x higher usage when using PyTorch. Data stru…
-
## 🚀 Feature
The LSTM layer in torch.nn should have the option to output the cell states of all time steps along with the hidden states of each time step.
## Motivation
When implementing Re…
-
不使用double policy的请看下,你的代码可以收敛么?我听别人说直接使用tanh再distribution 后sample会影响熵的计算,不知道为什么,可以问下么
-
Hi,
I'm trying to run your code in DVRL mode (according to the configurations you mention in the README file) and the results are significantly lower compared to the ones published in your paper (a…
-
I used the following two commands to identify broken links. `markdown-link-check` is https://github.com/tcort/markdown-link-check
``` bash
find ./Practical_RL/ -type f -name '*.ipynb' -exec jupyt…
-
https://datawhalechina.github.io/easy-rl/#/chapter7/chapter7
Description
-
Dear authors,
Thanks for the amazing work. Recently I followed the expert actions that I extract from `get_info()` function from the class `AlfredThorEnv`, however, the success rate is only slight…
-
### Search before asking
- [X] I searched the [issues](https://github.com/ray-project/ray/issues) and found no similar issues.
### Ray Component
RLlib
### Issue Severity
Medium: It contributes t…