-
Hi @bmazoure,
Your PPO +JAX implementation caught my eyes and this is a really cool repo!
Based on your [benchmark](https://wandb.ai/bmazoure/ppo_procgen_jax/reports/PPO-Procgen-JAX-version---V…
-
Hi, Great environment. Just wondering, is there a PPO baseline available for this environment?
-
## Problem Description
An error is reported when the file is runppo_atari.py
Run the ppo.py file correctly
error:gymnasium.error.NameNotFound: Environment `BreakoutNoFrameskip` doesn't exist.
…
-
Hi! This is nice work and it's easy but effective. I am wondering if you could open-source the PPO baseline model as well. I hope I can reproduce the results from Table 3 in the paper. It would be v…
-
Hi,
The current PPO implementation does not seem to account for time limits. While the `EpisodeWrapper` from brax is used, which tracks a truncation flag ([source](https://github.com/google/brax/bl…
-
Hi, thank you very much for providing the code. When I executed ppo_training_mh, I encountered the error code below.
The package versions I use are the same as yours except pytorch-geometric(2.5.3)
…
-
**Machine: MAX1100**
**ipex-llm: 2.1.0b20240421**
**bigdl-core-xe-21 2.5.0b20240421
bigdl-core-xe-esimd-21 2.5.0b20240421**
[Related PR](https://github.com/intel-analytics/ipex-llm…
-
Hi,
First of all thanks for interpax!
I see that interpax depends explicitly on numpy
-
Hi this a great project thankyou to share it , i transfert the code to ros2 humble and is working , now i change the algorithm to PPO but is not working can you give me some tips and tricks to implem…
-
## What
Add Curiosity driven exploration to PPO.
## Why
It's been shown [citation needed] that Curiosity improves agents' performance on sparse reward environments.