-
Dear Barhate,
Hi!!! Thank you for your sharing of the code very much! I can reproduce your results and they look super cool!
During my playing around,
may I consult in file PPO.py. line 83, …
-
Hi,
I have encountered error after kaggle submission. The following is error log from the game play in Kaggle. The game only plays for 1 turn and then stop. I used Python 3.7 to train the model
``…
-
I tried launching NLE using ray.rllib on Colab:
```
import gym
import nle
from nle.env.tasks import NetHackChallenge
import ray
from ray.rllib.agents.ppo.ppo_torch_policy import PPOTorchPolicy
…
-
@mugiwarakaizoku
I'm having trouble getting SAC to learn Cartpole effectively. Below is sample output of one of the better trials, but in most trials, it can't even break above a total reward of 1…
-
Is it possible to change the configuration (HFOV of sensors, POSITION of sensors, RADIUS, HEIGHT) of Agent_0 in the PointNav Baseline? When I try to change the base task config, either by altering the…
-
Hi, thanks for the awesome project.
I'm using following environment:
- stable-baselines3 0.8.0
- imitation 0.2.0
I modified examples/quickstart.py as following.
- Deleted BC and AIRL
- Modi…
ghost updated
2 years ago
-
**Important Note: We do not do technical support, nor consulting** and don't answer personal questions per email.
Please post your question on the [RL Discord](https://discord.com/invite/xhfNqQv), [R…
-
### What does it mean when we roll out PPO with numsteps > episode length
I know from the code that it will recycle the environment after you pass the terminal timestep. The question that I have is…
-
I think it's still TF experts right now (incompatible with our repo since torch port). Addresses part of #215.
-
## Purpose
The purpose of this issue(discussion) is to introduce a series of prs in the near future targeted to releasing tianshou's full benchmark for MuJoCo Gym task suite.
This benchmark will inc…