-
Hi, contributing this example of how to read the atari files directly, in case anyone wants to do that.
Note that the data is stored in the same temporal sequence it was logged, as you can see by w…
-
## 🚀 Feature
The LSTM layer in torch.nn should have the option to output the cell states of all time steps along with the hidden states of each time step.
## Motivation
When implementing Re…
-
## Motivation
Twin Q/ensemble Q functions are used in many RL algorithms and mitigate Q overestimation. My understanding is that TorchRL only deals with ensembles in the loss functions. This is fine …
-
### Search before asking
- [X] I searched the [issues](https://github.com/ray-project/ray/issues) and found no similar issues.
### Ray Component
RLlib
### Issue Severity
Medium: It contributes t…
-
-
Hi,
if I run the following script to train a model, one can aware significantly performance drawbacks in 1.2.0 than in 0.4.0.
```
import torch
import torch.nn as nn
import torch.optim as opt…
-
Hello,
I used Tainshou 0.5 on a custom environment running on a Windows PC. I was impressed by the training speed of the PPO agent, which exceeded 2000 iterations per second.
```python
import t…
-
I used the following two commands to identify broken links. `markdown-link-check` is https://github.com/tcort/markdown-link-check
``` bash
find ./Practical_RL/ -type f -name '*.ipynb' -exec jupyt…
-
https://datawhalechina.github.io/easy-rl/#/chapter7/chapter7
Description
-
Thank you for your excellent work.
However, when I train multiple agents in a highway environment, the network doesn't seem to converge.
The checkpoint rewards obtained after training for every thou…