-
Hi,
When I tried to run `python3 sb3_train.py`, there were several errors in the original code.
- The `env_name` here should have two underscores between the environment name and suffix "open", …
-
Is there any setting can control the GPU's max memory usage, seem it will just directly use up as much as possible. When GPU memory is full, just quit.
Or more GPUs are needed. Any suggestion, than…
-
Great job! I like this comprehensive benchmark
I just developed [CoPO, a MARL algorithm explicitly modeling the coordination between self-interested agents](https://decisionforce.github.io/CoPO/), …
-
Hi all,
I was wondering if the PPO-based MARL algorithms you use in the paper are taken from RLlib or whether they are already available in the library without the need of an RLlib interface.
I …
-
Hello, I would like to ask, since it is a scene of homogeneous agents, why does the shared actor need to use ID as a condition? Is it ok if you don't use the ID as part of the state?
-
### Required prerequisites
- [X] I have read the documentation .
- [X] I have searched the [Issue Tracker](https://github.com/PKU-MARL/omnisafe/issues) and [Discussions](https://github.com/PKU-MARL/o…
-
Hello!
Sorry to bother you. Thanks for this brilliant paper and all the work. I find the idea of applying MARL to distribution network is very enlightening and try to reproduce it.
But when I …
-
### The problem
I am using the OctoPrint (OctoPi on Rpi3) with Klipper.
I have configured Long running commands according this picture.
![image](https://user-images.githubusercontent.com/16965519…
-
Hi, is the `step()` funtion in `warp_drive` individual for each thread agent? In Figure 1 displayed in the paper, it seemes like that each thread agent in the block maintains a individual `step` funct…
-
Dear,
what is the difference between Safexp-PointGoal1-v0 vs SafetyPointGoal1-v0?
I find that, result from SafetyPointGoal1-v0 is generally smaller than Safexp-PointGoal1-v0