-
I am conducting reinforcement learning for a robot using rsl_rl and isaac lab. While it works fine with simple settings, when I switch to more complex settings (such as Domain Randomization), the foll…
-
https://github.com/leggedrobotics/rsl_rl/blob/master/rsl_rl/modules/actor_critic.py#L121
The action noise had better use positive parameterization, e.g.
self.distribution = Normal(mean, torch.ones_…
-
Hi, I am new to tianshou and RL. I created a env and used ppo in tianshou to run. But I found the action sampling is out of range. So I searched for, and I found map_action. But it seem not used in tr…
-
I am conducting reinforcement learning for a robot using rsl_rl and isaac lab. While it works fine with simple settings, when I switch to more complex settings (such as Domain Randomization), the foll…
-
@Kismuz,
I believe I have encountered a framework (A3C) limitation.
While training a few of my recent models I noticed a strange behavior. For the first part of training everything seems to work fi…
-
I think you should use `tf.stop_gradient()` in https://github.com/coreylynch/async-rl/blob/master/a3c.py#L164. Otherwise, after some training the policy tends to use one action exclusively. Took me a …
-
-
Hi team getting the following error while enabling 4-bit and LORA
```
File "/root/miniconda3/envs/open/lib/python3.11/site-packages/deepspeed/runtime/engine.py", line 262, in __init__
self._c…
-
-