-
As I've been taking lots of notes while reading papers related to Rainbow, I thought I'd set up the documentation website and flesh it out gradually. I'll link a pull request with a first version of t…
-
### What happened + What you expected to happen
The method `convert_to_torch_tensor` fails and returns the following TypeError:
```
TypeError: can't convert np.ndarray of type numpy.str_. The onl…
-
In many on-policy RL algorithms, we would like to pause the environment while we synchronously wait for the update of the weights. Or, in other RL game environments, the game timer only steps every ti…
-
Hi, this is a nice project for hybrid action space, and I see you mentioned PDQN/HPPO in `README.md`. Do you have some experiment results about these algorithms in this environment. If not, we want to…
-
I'm interested in the condition control part implemented by RL, but I got a confusion about the RL training. Does the training of RL on a set of training data or just one image online? From what I saw…
-
Hi everyone,
It has been cool to see the recent flurry of contributions to this package, especially by @jeremiahpslewis. In a [recent discussion](https://github.com/JuliaReinforcementLearning/Reinf…
-
**Thank you for your contribution !**
I want to know in all of the self-attention RL algorithms, which one has the best performance?
Thank you !
-
- [ ] I have marked all applicable categories:
+ [ ] exception-raising bug
+ [ ] RL algorithm bug
+ [ ] documentation request (i.e. "X is missing from the documentation.")
+ [ ] ne…
-
Dear author, I am implementing the Multi-agent settings using the Highway-v0. I am not able to achieve stable training and the vehicles can run off the roads without terminating the environment. I too…
-
### What happened + What you expected to happen
I converted existing code working on 2.7 to 2.20 (new api)
The error:
File "/opt/project/trading/training/model/rl/multi_agent/ppo/equity/trainer…