-
### 🐛 Bug
Running Recurrent PPO on CartPole in a background notebook in Kaggle after 6 hours the task crashed before finishing
### To Reproduce
It was a simple test on cartpole environment. Here th…
-
Hi Everyone,
I've been using ppo-tfjs for the last month and find it to be an incredible library, thank you so much for making it! I've been working on a fork over at https://github.com/alistairhea…
-
## A question
- I want to build a small scale version of **Open AI Five**
- And I learnt that it uses LSTM + PPO
- suppose I build a network model using LSTM, then should I use this network for …
-
Hi, Great environment. Just wondering, is there a PPO baseline available for this environment?
-
When running the PPO baseline on my M1 Mac using the command `python ppo.py --save_policy`, I encounter `ValueError: Unrecognized name format` during the policy-saving process within the _save_network…
-
Hi @bmazoure,
Your PPO +JAX implementation caught my eyes and this is a really cool repo!
Based on your [benchmark](https://wandb.ai/bmazoure/ppo_procgen_jax/reports/PPO-Procgen-JAX-version---V…
-
Release test **rllib_learning_tests_pong_ppo_torch.aws** failed. See https://buildkite.com/ray-project/release/builds/16725#018fe1f2-a6ac-4002-b08b-6d5c34f87e40 for more details.
Managed by OSS Test …
-
train.py 和config_ppo.yaml 中low_level_load_path是如何生成的
evaluate.py中设置了lower_model和upper_model
报错Encoder type cnn not supported!
4个upper_model都试过了加载后的encoder_type是cnn而非pixel
有更详细的训练或者验证介绍嘛
-
I was running the script from step3: python3 train.py --step 3 --deployment-type single_gpu
The training.log shows this:
A decoder-only architecture is being used, but right-padding was detected! …
-
## What
Add Curiosity driven exploration to PPO.
## Why
It's been shown [citation needed] that Curiosity improves agents' performance on sparse reward environments.