-
### ❓ Question
Hello, If I want to plot training rewards with different algorithms about one Env,
such as `:
python scripts/plot_train.py -a td3 sac ddpg -e PandaReach -f logs/ -w 500 -x steps`
b…
-
Hi @thomascent,
I've been trying to use your envs with a stable_baselines algo (here's the cleaned-up [repository](https://github.com/MartinaRuocco/diy-gym/tree/master/examples/RL_test)) but I had…
-
Dear CORL Team,
Firstly, I would like to express my appreciation for your work on the CORL codebase. The clean, single-file implementation coupled with a robust performance report has greatly impre…
-
Hi, in PyTorch docs we can read:
```
torch.nn only supports mini-batches. The entire torch.nn package only supports inputs that are a mini-batch of samples, and not a single sample.
For example, …
-
HI,How to ues in python3 , i have failed used in python3 . because ImportError: dynamic module does not define module export function (PyInit__tf2),can you tell me,thanks.
GXJll updated
3 months ago
-
开发者您好,我想请问本项目是否可以调用其他模型的接口进行对比试验呢?
-
I'm trying to use a DDPG agent with actor and critic networks, and a TFUniform replay buffer, training on my custom environment.
I've extracted a training experience from the buffer using:
```
da…
-
Hello,
Nice project =)
I created a colab notebook to try it online directly: https://colab.research.google.com/drive/19bdAiKZY0r5OR3gEv7164CjDOdMRGYqt
Btw, why didn't you use `deterministic=…
-
Does anyone have recommendations on what to do to fix this? My model essentially learns to just buy and hold stocks instead of exploring trading strategies. My learning rate spread is quite large (bet…
-
Disclaimer: I am not completely sure if this is a bug of PFRL.
When I ran SAC, and TD3 on my university's cluster without a GPU, I observed that memory usage gradually increased and finally reached…