-
On executing trpo_continous.py, I get the following error:
> [2017-07-01 23:52:58,375] Making new env: CartPole-v0
> [TL] InputLayer continous_shared/continous_input_layer: (?, 3)
> [TL…
-
In train.py, I see a central agent,SL agent and RL agents. They are running in different CPU cores with multiprocessing package. And RL agents get the weights of policy and value network from central …
-
Hello, I'm trying to port RL maddpg algorithm on smaug. Is there any documentation you can provide to follow for integrating the new algorithm?
Thanks
-
Hi,
I would like to ask whether there is a jax-based code.
And whether there are some recommendations about jax-based offline rl algorithms.
Thanks!
-
### What happened + What you expected to happen
I am having issues loading a DreamerV3 checkpoint for inference. Similar to what was discussed in #40312, I assume it has to do with the old/new API.
…
-
I am interested in using Flow for VANETs (Vehicular Ad hoc NETworks) routing protocols, which play a key role in the design and development of Intelligent Transportation Systems. Besides RL,
genetic…
-
# What has been done:
- Machine Learning course:
- Week 5 ✔️
- Week 6 ⌚
- EURO2022 - abstract ⌚
- Reinforcement Learning:
- Restructure code
- Plot some form mean policy next to…
-
The paper "gCastle: A Python Toolbox for Causal Discovery" claims that "gCastle includes ... with **optional GPU acceleration**". However, I don't know how GPU acceleration can be used on this package…
-
# What has been done:
- Gregynog ✔️
- Booking flights/hotels/registrations for conferences ⌚
- Julia course ✔️
- Fixed Reinforcement Learning algorithm bug ✔️
- Output of RL algorithm 💯🎉
# To …
-
My understanding is that most RL algorithms will focus on supporting gymnasium going forward and that will be the standard. Trying to get ray rllib or other RL libraries with gym environments is prett…