-
### What happened + What you expected to happen
Hello together,
trying to run the action_masking_example under `rllib/examples/rl_modules/action_masking_rl_module.py` after commit https://github.…
-
### Required prerequisites
- [X] I have read the documentation .
- [X] I have searched the [Issue Tracker](https://github.com/PKU-Alignment/safe-rlhf/issues) and [Discussions](https://github.com/PKU-…
-
### What happened + What you expected to happen
The problem is about numpy version when I do checkpoints restore on model. On numpy>=2.0.0 works, but on numpy=1.20, but it works only >=2.0.0. I can n…
-
Hi @praveen-palanisamy
I have been working on macad-gym successfully over the past few months using PPO and many other algorithms. Now I am trying to use DDPG using RLlib which requires continuous…
-
# 🚀 Feature Request
Part of #42. Depends on #44. Once an environment is set up, it will be easy to train several of the RL algorithms provided by `pytorch`. All of these algorithms should b…
ghost updated
5 years ago
-
I want to build an RL algo that will understand the concept of beating a benchmark (say S&P500), at a tic level. So if a tic is constantly beating the benchmark, the algo should prefer to pick that ti…
-
# Context
Hi, I like what has been done with RL-ViGen and I like that this benchmark exposes well that most approaches are not perfect in all scenarios. I think ideally, RL-ViGen should be used for …
-
**Description**:
The RL and IRL algorithms need tuning to perform well (especially the Adversarial ones). We need to put some time and tune them and see if they can perform well if we want to use the…
-
This issue serves for informing about and discussing the next major release of Tianshou, after which the library can be considered mature and stable from our perspective. The progress and the related …
-
I am conducting reinforcement learning for a robot using rsl_rl and isaac lab. While it works fine with simple settings, when I switch to more complex settings (such as Domain Randomization), the foll…