-
**Description**
There seem to be some edge cases around identity functions that cause problems for [SE-0347](https://github.com/apple/swift-evolution/blob/main/proposals/0347-type-inference-from-de…
-
I am looking for a simple reinforce/a2c example to work with Mazebase environments.
It seems like Mazebase repo is no longer compatible with this repo?
I get this error
from mazebase.torch_f…
-
I would like to know how to be able to train or test your code?
When I try to run SAFE_RL.ipynb, it prompts me with
FileNotFoundError: [Errno 2] No such file or directory: 'teacher_models/A2C_pong_t…
-
In train.py, I see a central agent,SL agent and RL agents. They are running in different CPU cores with multiprocessing package. And RL agents get the weights of policy and value network from central …
-
DQN is also not training but that'll be addressed after DQN is restructured.
-
https://github.com/openai/baselines/blob/9ee399f5b20cd70ac0a871927a6cf043b478193f/baselines/a2c/a2c.py#L100
Wondering what these masks are for, where they are applied and which class they are a mem…
-
Algorithms in ```qpolgrad``` have been organized to define functions for loss calculation. Those functions are then called in the ```update``` function for the algorithm. A2C and PPO need to be brough…
-
Versions:
uniforms: ^3.10.1
uniforms-mui: ^3.10.1
```
type A1 = {
id_type: 'a1';
id: string;
};
type A2 = {
id_type: 'a2';
id: string;
};
type C = {
name: string;
addres…
-
Tried to run rl with new gym wrapper code and it gives out the following error
`expected dense_input to have shape (1, 10) but got array with shape (1, 2)`
this is the code `import asyncio
from…
-
Hello,
There is a pre-trained agent (agent_a2c.zip) that is included in Stock_NeurIPS2018.
I also reran the notebooks 1 and 2 to train agent_a2c. However the backtest result of the new agent is mu…