-
Hi,
I want to reproduce the comparison between A3C and GA3C in Table 2 in your paper.
I wonder if the A3C experiment can be done using this repo?
Thanks
hma02 updated
7 years ago
-
I am kind of confused of the ensure_shared_grads here https://github.com/ikostrikov/pytorch-a3c/blob/master/train.py#L13. Here, the `grad` is synced only when it is `None`. I think we need to set `sha…
-
https://github.com/dacthai2807/odmc-wrsn/blob/cd22ebadb67f79b2f743439b97fe9a30b85de4c5/main_multiprocessing.py#L164
Ko nên truyền train_data, valid_data vào nữa em. Vào trong thì tự sinh data sau.
-
Hi, I'd to have a question about the following block
https://github.com/dgriff777/rl_a3c_pytorch/blob/eb5c9b909abc02911b45e325f7a7c619d3b0fa46/test.py#L60
```
if player.done and not player.…
-
Hi there,
Thanks for the implementation!
I tried running the code to solve the breakout game(v0). However, after 3 days of training the results were quite far from one a DQN would obtain - the a…
-
### What happened + What you expected to happen
I am a Mac M1 Pro user. I have run RL jobs successfully with both TensorFlow and PyTorch before this. But I wanted to check features provided by this f…
-
-
**### I run the Command: python3 pensieve_torch.py --model_type=1**
File "/Library/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/process.py", line 315, in _bootstrap
…
-
Do you have a version of the code in Python 3.x + TensorFlow 2.x? This'll help me run on a platform that does not have Python 2.7 + TensorFlow 1.1.0.
-
The loss may be negative number in the model. The reason is that the reinforce loss is often to be a negative number since the reward is the larger the better. But I am very confusing about how negati…