-
This issue serves for informing about and discussing the next major release of Tianshou, after which the library can be considered mature and stable from our perspective. The progress and the related …
-
First of all, thank you for open-sourcing this algorithm.
I am trying to train a quadruped robot locomotion policy with multimodal input, including egocentric depth vision, on complex terrain. Howev…
-
# 🚀 Feature Request
Part of #42. Depends on #44. Once an environment is set up, it will be easy to train several of the RL algorithms provided by `pytorch`. All of these algorithms should b…
ghost updated
5 years ago
-
**Description**:
The RL and IRL algorithms need tuning to perform well (especially the Adversarial ones). We need to put some time and tune them and see if they can perform well if we want to use the…
-
- [X] I have marked all applicable categories:
+ [ ] exception-raising bug
+ [ ] RL algorithm bug
+ [ ] system worker bug
+ [ ] system utils bug
+ [X] code design/refactor
…
-
Hello,
Thanks for this great library.
I have a question. I want to use RL2 and specifically the RL2TRPO algorithm with discrete action-space. However, it seems that the current implementation do…
-
I have been using these two routines to figure out the best learning rate to apply with awesome results on SAC. However, the changes in the `temperature` alter those values along the way. Probably wou…
-
### What happened + What you expected to happen
As per my best knowledge, the repro script covers the version of `action_masking_rlm.py` example file which was shipped in release 2.34. However, I adj…
-
- [ ] I have marked all applicable categories:
+ [ ] exception-raising bug
+ [ ] RL algorithm bug
+ [ ] documentation request (i.e. "X is missing from the documentation.")
+ [ ] ne…
-
### What happened + What you expected to happen
When using the `on_episode_end` callback, the environment is reset before the callback is called. This means that accessing internal environment vari…