-
## Problem
Hello!
Thanks so much for sharing the code! There is a question.
I want to load the trained learner net and the trained reward_net to train learner and reward_net again.
So how d…
-
Mackenzie and Gail have noted that, at least in the wet-lab, only a single study at a time has been test-plated. Before full-launch, we should test two or more studies being plated at the same time, a…
-
Nice, finally found this project that updates the policy using reward from discriminator and aligns itself with the algorithm in the GAIL paper. In many other libraries, they just use the reward from …
-
Hi, I am the developer of [DI-engine](https://github.com/opendilab/DI-engine), and we plan to add an inverse RL/GAIL related algorithms benchmark. Your project has detailed implementation and visualiz…
-
Broke this request out from #381 because it is adding a new functionality to the prep sheet, not reproducing the functionality of the existing wet-lab system, so it falls outside the scope of soft lau…
-
Hi,
I manually controlled the ant robot from this example (https://github.com/NVIDIA-Omniverse/IsaacGymEnvs/blob/main/isaacgymenvs/tasks/ant.py) and recorded the corresponding joint angle values.
I…
-
[//]: # (Copyright Siemens AG, 2021. Part of the SW360 Portal Project)
[//]: # (This program and the accompanying materials are made)
[//]: # (available under the terms of the Eclipse Public License…
-
**High Level Description**
When replaying the .jsonl file with large amount of (~30) self-controlled vehicles, the envision video looks laggy.
**Desired SMARTS version**
0.5.1
**Problems**
Th…
-
I am still struggling with the implementation of a recurrent policy. The trick from [#1](https://github.com/Khrylx/PyTorch-RL/issues/1) worked and I can now start running my RNN GAIL Network. But no m…
-
To improve stability and robustness of policy, implement proximal policy optimization (PPO):
- https://arxiv.org/abs/1707.06347
- code: https://github.com/ikostrikov/pytorch-a2c-ppo-acktr-gail