-
Hello,
I'm working on training a gail agent with human expert demonstrations dataset (provided by robomimic repository - Lift task in robosuite simulator).
I modified this dataset a little bit to …
-
Hi, var discr not found while I am trying to run train_ppo.py
![image](https://user-images.githubusercontent.com/8272997/195036154-a92a9540-43f8-41d3-8811-ac1361a9182e.png)
-
I'm trying to apply GAIL using retroEnv "SonicTheHedgehog-Genesis" but I'm getting some errors. Apparently, the env is not recognized. Does anyone have any idea what could be causing this?
below th…
-
**Is your feature request related to a problem? Please describe.**
(not a real story)
Let's admit I've got a great model that took weeks to train. It has 10 hidden layers of 1024 units.
Inference w…
-
## Bug description
Description of what the bug is.
## Steps to reproduce
Code or a description of how to reproduce the bug.
## Environment
- Operating system and version:
- Python version:
…
-
````
def calculate_log_pi(log_stds, noises, actions):
gaussian_log_probs = (-0.5 * noises.pow(2) - log_stds).sum(
dim=-1, keepdim=True) - 0.5 * math.log(2 * math.pi) * log_stds.size(-1)…
-
while some of the library is configured to allow cpu usage, it seems like not all parts are.
if trying to use habitat_basemap_location=self.device, I don't have a gpu to use but am using habitat-ap…
-
## Bug description
I try to rollout the trajectories by PPO/A2C for task "BreakoutNoFrameskip-v4".
An error occurred "ValueError: Observation spaces do not match". I think the Observation got trans…
-
Hi,
This is an agent learning to drive and keep in lane
Red = PPO
Grey = PPO+GAIL @ 0.5 reward strength
Demonstrations recorded by PPO (red) model with deterministic inference
Any suggestions as …
MrOCW updated
2 years ago
-
I had no problem training the agent days ago but now after a lot of changes to the env. It crashes everytime i try to train it. Don't know why it suddenly happens? It works fine in heuristic but it do…