-
Hi,
thanks for the amazing work of RL environments using JAX. I was wondering if you have any plans to write Actor-Critic agents for this work?
-
If I understand the current PPO code correctly, this instantiates completely separate actor and critic models, without any layers shared between them. (But correct me in case that is wrong?)
Instea…
-
Reduce duplication of similar Actors/Critics with only the hidden layers being different - generally improve the readability of the code for creating the networks.
-
Stock Dimension: 30, State Space: 2371
{'batch_size': 64, 'buffer_size': 100000, 'learning_rate': 0.001, 'learning_starts': 100, 'ent_coef': 'auto_0.1'}
Using cpu device
Logging to /content/drive…
-
### What happened + What you expected to happen
# What happened
Using the `Algorithm.add_module` with a `module_state` does not use the module state, but instead loads or builds the module directly…
-
The current implementation of `ActorCriticBase` makes it a bit trick to have custom actor and critic networks that have shared layers. This is because the instantiation of the networks happen in the `…
-
Fairseq contains many NMT models but models with Reinforcement Learning are absent.
It would be great if that is added
-
Mr Siraskar,
Hello ! I have read your paper and your code, which are very helpful to me, thank you for sharing your valuable work to us !
I have a question :
Why you designed the actor network …
-
-
Base line run of Rebrac on half cheetah medium v2
https://wandb.ai/jnqian/CORL/runs/a4876f1d-be93-4616-b5d8-2ec84a1a9f5a