-
I am trying to spawn mulitple habitat enviroments parallely. [Threads](https://github.com/pushkalkatara/visual-navigation-agent-pytorch/blob/b8d5b40313b1c4d974d26c8a1693a2e2f17418de/agent/training_thr…
-
## 🚀 Feature
Implement a dataloading functionality for reinforcement learning state, action pairs, with assigned policy scores, transitional probabilities and rewards.
Implement a set of gradient al…
-
### Describe the problem
Currently we only have a simple pytorch example for A3C. It should be possible to "port" a bunch of pytorch algorithms onto the TorchPolicyGraph class and get them basicall…
ericl updated
4 years ago
-
I've seen this multiple times, never before with the current ray backend though.
```
+ docker run -e RAY_USE_XRAY=1 --rm --shm-size=10G --memory=10G cfca7c87523a7d361ba790a1c243ee28904b7148b9b0bd0…
ericl updated
4 years ago
-
I noticed that for the vectorized environments, all of environment will take a step and will wait for all the steps to finish before starting to take another step. The issue with my customized environ…
-
Tensorflow is hard to debug and hard to teach with. We are rolling out an experimental course version using PyTorch.
The following files will have to be ported:
- [x] `week1_intro/primer/recap_t…
dniku updated
4 years ago
-
### System information
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: Linux Ubuntu 18.04
- **Ray installed from (source or binary)**: binary (via pip)
- **Ray version**: 0.7.6
- …
-
### System information
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: docker Ubuntu 16.04 image
- **Ray installed from (source or binary)**: pip
- **Ray version**: 0.5.3
- **Py…
-
Hi, thanks for the great implementation. I learned a lot about PPO by reading your code.
I have one question regarding the state_values.detach() when updating PPO.
```
# Finding Surrogate Loss:
…
-
### Tensorflow fails initializing a trainer
I use rllib and have tensorflow (1.1.4) and pytorch installed in a python 3.6.9 environment. When initializing a trainer `trainer = ppo.PPOTrainer(env=…