-
Hi, thank you for your work, it's amazing! I'm a student who just started DRL. I set up the simulation environment according to the tutorial and used your original program to train (by executing 'pyth…
-
### 🚀 Feature
Stochastic Weight Averaging (SWA) is a recently proposed technique can potentially help improve training stability in DRL. There is now a new implementation in `torchcontrib`. Quoting/p…
-
Excuse me, Is there any method that does not require network mode training?
Because I think the network communication time may affect the execution speed of each step in RL and thus affect the traini…
-
Hey Jingliang, just saw your paper for the first version of DSAC.
It is impressive and clear. I am quite interested in the implementation of TD4, which is quite interesting in my view.
Would you …
-
Sorry, I'm here to ask a question again
I'm trying to execute GDAM.py
i can't find the file
OSError: File /home/wenzhi/GDAE/Code/assets/launch/multi_robot_scenario.launch does not exist
![2023-0…
-
On running the lagrangian version of SAC I get the following curve for costs. I tried changing the constraint limits to a range of values and didn't get much benefit:
![lagrangian_sac_pointgoal1](h…
-
SAC algorithm in PyTorch implementation has a serious bug
`q_params = itertools.chain(ac.q1.parameters(), ac.q2.parameters())`
`itertools.chain` will become empty after the first iteration, so e…
-
Implement the best practices from multi-agent Rl community and stablebaselines3 into our algorithm. Further analyse similarities between petting zoo multi-agent implementation to current RL implementa…
-
I've tried to make the environment work with the baselines fork stable_baselines (https://github.com/hill-a/stable-baselines). It runs, but the results shown when I'm running plot_energyplus is always…
-
不使用double policy的请看下,你的代码可以收敛么?我听别人说直接使用tanh再distribution 后sample会影响熵的计算,不知道为什么,可以问下么