-
Hi, I was trying to use A2C on the ingolstadt21 map, but when I set the ```single_agent=False```. I'd receive this error.
```
Traceback (most recent call last):
File "/home/ubuntu/Videos/test.p…
-
Dear authors,
Thanks for building such ambitious environments for MARL research. In your paper, I found it will take 10^9 steps to run the simulation for each agent. In order to train agents, how m…
-
I just set up this repository following the README docs. The evaluation scripts for Scene Priors and Non-Adaptive A3C worked, but not for SAVN. Because pytorch 0.4.1 is out of date, I tried with 1.8.1…
-
### Search before asking
- [X] I searched the [issues](https://github.com/ray-project/ray/issues) and found no similar issues.
### Ray Component
RLlib
### What happened + What you expect…
-
Searching for ["ray pytorch"](https://www.google.com/search?q=ray+pytorch&oq=ray+pytorch) leads to [this page](https://docs.ray.io/en/latest/using-ray-with-pytorch.html)
We should update this page…
-
### What is the problem?
The recently added torch implementation of PPO #6826 is over 5X slower when training on atari (breakout) and also ends up slowly consuming all the system RAM (perf/ram_…
-
## 🐛 Bug
Calling xm.spawn() with nproc=8 makes colab notebook running indefinitely.
it works with nproc=1
If i stop the running cell and do !ps -ef | grep python, there are like 30+ python proc…
-
Stable_baselines3 can only train the env of "gym"? Or how to convert the magent-type env to gym-type env that can be trained by stable_baselines3?
-
## 🐛 Bug
It seems like there is very small memory leak during forward and backward propagation through network that can lead to memory overflow after many hours of training.
I stumbled upon this…
-
### What is the problem?
I found multiple runs for PPO still have different performance even we set the same seed.
How can we obtain the exact same result with the same seed?
Current strate…