-
I tried running the PPO2 example using the breakout game on my MacBook, without modifying anything in the scripts or configurations and I am only able to get up to 19.6 score. Why is that? Is there an…
-
Hi this a great project thankyou to share it , i transfert the code to ros2 humble and is working , now i change the algorithm to PPO but is not working can you give me some tips and tricks to implem…
-
### What happened + What you expected to happen
I can’t seem to replicate the original [PPO](https://arxiv.org/pdf/1707.06347) algorithm's performance when using RLlib's PPO implementation. The hyp…
-
Release test **long_running_many_ppo.aws** failed. See https://buildkite.com/ray-project/release/builds/17769#019028f5-f349-483a-8645-b6529e00dc9a for more details.
Managed by OSS Test Policy
-
### What happened + What you expected to happen
Hello,
I am encountering an error while training a PPO agent using RLlib. During training, I receive the following error message:
`File "/opt/conda…
-
### 🚀 Feature
Hello guys,
After watching this video :
[https://www.youtube.com/watch?v=WoLlZLdoEQk](url)
I had the idea to extend the NatureCNN to NatureCTN1D this way :
```
class Chomp1d(nn…
-
您好,感谢您的工作!
我在运行PPO算法时,其中PPOGitHub - model.py - 337行“state = t.from_numpy(state).to(self.device)”报错,
TypeError: expected np.ndarray (got tuple)
请问是什么问题呢?谢谢您
AI-HM updated
11 months ago
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
[2024-06-07 10:17:14,980] [INFO] [real_accelerator.py:191:get_accelerator] Setting ds_accelerator t…
-
Here is a Robot Script for exporting the PPO_reasoned_merged.owl file:
```
./robot filter --input ppo.owl --term PPO:0002300 --select "self annotations descendants" --signature true export --he…
-
They are very complicated, and I don't think they are necessary for data ingest. I'm going to try it, and @jdeck88 can let me know if it messes anything up.
For example:
PPO:0002356 -- abscised …