-
## Describe the bug
Not quite sure if this is supported behavior, but if I set `functional=True` for the A2C loss and `shifted=True` for `TD0Estimator` I get an internal error.
## To Reproduce
…
-
Hi @eleurent . thank you so much for the contribution. Please I need to know how you figured out the hyperparameters of DQN in the highway run env. did you use optuna for optimizing the hyperparameter…
-
```ts
type A1 = {
id_type: 'a1';
id: string;
};
type A2 = {
id_type: 'a2';
id: string;
};
type C = {
name: string;
address: string;
}
type A1c = A1 & C
type A2c = A2 & C
…
-
### ❓ Question
Hello,
I first optimize A2C on 1mln steps using RL Baselines3 Zoo:
Firstly i have changed `a2c.yml` in RL Baselines3 Zoo to work with RAM version of Seaquest:
```
atari:
pol…
-
Lets implement that shiz yo
-
On execute this code:
````
df_summary = ensemble_agent.run_ensemble_strategy(A2C_model_kwargs,
PPO_model_kwargs,
…
-
-
Versions:
uniforms: ^3.10.1
uniforms-mui: ^3.10.1
```
type A1 = {
id_type: 'a1';
id: string;
};
type A2 = {
id_type: 'a2';
id: string;
};
type C = {
name: string;
addres…
-
https://github.com/AI4Finance-Foundation/FinRL-Tutorials/blob/master/4-Optimization/FinRL_HyperparameterTuning_using_Optuna_basic.ipynb
I am getting the following error while recreating the above n…
-
Write a blog about the visualization of A2C playing Atari Pong. It seems that many actions are about the same most of the time (the horizon is limited by gamma) and rarely specific actions are _intend…