-
Implement the best practices from multi-agent Rl community and stablebaselines3 into our algorithm. Further analyse similarities between petting zoo multi-agent implementation to current RL implementa…
-
How can I make my own dataset for mask RCNN training ?
-
- [ ] I have marked all applicable categories:
+ [ ] exception-raising bug
+ [ ] RL algorithm bug
+ [ ] documentation request (i.e. "X is missing from the documentation.")
+ [x] ne…
-
I was trying to run [Learning to Drive in 5 minutes](https://github.com/araffin/learning-to-drive-in-5-minutes?tab=readme-ov-file#quick-start) but the Stable Baselines version is dependent on TensorFl…
-
**Important Note: We do not do technical support, nor consulting** and don't answer personal questions per email.
**Describe the bug**
When I use the available SAC agent for AntBulletEnv-v0 to cre…
-
Can anyone suggest hyperparams for A2C when training with the parking-env from the [highway-env](https://github.com/eleurent/highway-env/blob/master/docs/source/index.rst) repo?
-
### Expected results
In the caffe2 implementation, there is a field TRAIN.WEIGHTS which allows us to use models pretrained on datasets such as coco. Can I merely use pretrained resnet backbones b…
-
Want to do hand/body tracking with detectron2
-
The policy that I have tried is DDPG and SAC. I used master branch and below is the two command to reproduce the error.
`python train.py --algo sac -n 5000
`
`python train.py --algo ddpg -n 5000
`…
-
Dear @pzhokhov @matthiasplappert @christopherhesse et al.,
Thank you for providing an implementation of DDPG. However, I have been unable to get it to learn well on the standard MuJoCo environmen…