-
Hello,
You advertise tianshou as being fast and provide in the readme a comparison table.
However, no reference code is linked to reproduce the results.
So, I decided to create a colab notebook…
-
In QR's paper, it claims **Distributional reinforcement learning (Jaquette, 1973; Sobel, 1982; White, 1988; Morimura et al., 2010b; Bellemare et al., 2017) focuses on the intrinsic randomness of retur…
-
Hey man, I came across the need to have a route planner, because I have several intersection points drawn on the map, but none of these points has an optimized route, is there something already implem…
-
Paper: https://arxiv.org/abs/1710.10044
@ku2482 Please tell us if you want to work on that ;)
-
https://github.com/araffin/learning-to-drive-in-5-minutes/blob/ccb27e66d593d6036fc1076dcec80f74a3f5e239/algos/custom_ppo2.py#L165
Hi, I think the logic of updating mb_rewards is wrong here.
for ex…
-
Hi, thank you for your great work!!
I'm interested in contributing to Stable-Baselines3.
I want to implement SAC-Discrete([paper](https://arxiv.org/abs/1910.07207), [my implementation](https://git…
-
For Examples as a use of saying hi to you, of course tbh.
```
Before (Must require abcdefghijklmnopqrstuvwxyz (LC and/or UC) to 0123456789 and !"#$&'()*,-.;:@\/%+?[]^_`{|}~ in order to do all this):…
-
[question] Is there any reason why DDPG and SAC don't have the implementation for discrete action space? And will appreciate it there are any suggestions for applying the DDPG with continuous action …
-
hello? is this relates to ERDL, ER-DQN, or the deepmind paper "statistics and samples in distributional reinforcement learning"?
if so, i'm quite curious, how to calculate the equation 7?
what I can…
ddlau updated
3 years ago
-
I'm normally against implementing very recent papers before they prove to be valuable but I would like to make an exception for that one, especially because of the good results. It was recently accept…