-
Dear Mr.hongzi
I was interested in your resource scheduling method. Now, I stuck in your network class. I can't understand why you used the blow function:
`loss = T.log(prob_act[T.arange(N), actions…
-
AGENT NAME: A3C
1.1: A3C
TITLE CartPole
layer info [20, 10, [2, 1]]
layer info [20, 10, [2, 1]]
{'learning_rate': 0.005, 'linear_hidden_units': [20, 10], 'final_layer_activation': ['SOFTMAX', …
-
I need to get a copy of `shared` neural network of type `torch::nn::Sequential`. It seems that there is no available API for this purpose at the moment. It seems that declaring and instantiating the n…
-
After value, logit, (hx, cx) = model((Variable(state.unsqueeze(0)),(hx, cx))) in train.py, the program doesn't go on. Do you have any idea?
-
Hi, when I try to run this code using CUDA I always am getting this error. It works fine on CPU and other programs work fine using CUDA. I tried to search online but I didn't find anything of use.
He…
-
Hi all, I run a3c in Ant-v1 and got this warning:
```
WARNING: Nan, Inf or huge value in QACC at DOF 0. The simulation is unstable. Time = 0.0000.
```
Are there anyone know what's the problem wi…
-
## 🚀 Feature
The RL implementations added do not have the num_workers option. I have a feeling this is because the code doesn't support a shared replay buffer.
### Motivation
Adding this would e…
-
Is this project active? (I don't see any other way to message Josiah.) I've been thinking of working on something similar but would rather contribute to an existing project than start from scratch. Bu…
-
**/home/account/anaconda3/envs/RL17/bin/python /home/account/Documents/Deep_RL_Implementations/results/Cart_Pole.py
/home/account/anaconda3/envs/RL17/lib/python3.7/site-packages/gym/envs/registration…
-
Hello !
I have a trouble with Gym-mupen64plus environnement and i don't know what to do.
When i start my project with Docker Container, the game doesn't go in Times Trials mode after the env initial…