-
I keep getting this error due to some in place changes to the variable a in sample_multiple:
`[W python_anomaly_mode.cpp:60] Warning: Error detected in AddmmBackward. Traceback of forward call that…
-
- [X] I have marked all applicable categories:
+ [ ] exception-raising bug
+ [ ] RL algorithm bug
+ [ ] documentation request (i.e. "X is missing from the documentation.")
+ [X] ne…
-
Hi,
I tried the following command to test the Humanoid-v2 task
`python train_rpp.py --env_name=Humanoid-v2 --save_dir=./tmp/rpp --rpp_value=False`
However, it didn't work and raised those error mes…
-
Hello,
I’ve been using the official recommended script as follows:
```bash
# dc
python3 experiment.py --env hopper --dataset medium --model_type dc --K 8 --embed_dim 256 --learning_rate 0.0001…
-
**Is your feature request related to a problem? Please describe.**
The current `pip install d3rlpy` installs a bunch of new packages and upgrades existing packages without user's consent. This terrib…
-
Overall, we should try to focus our efforts towards what's necessary for the paper.
1. DataDistance vs. ScoreMatching - do we also want to show that optimal control w/ data distance penalty is emp…
-
# Summary
1. There are issues with the scoring calculation of expert strategies in the maze2d environment.
2. The incorrect scoring calculation is a result of the expert strategies not being called …
-
**Is your feature request related to a problem? Please describe.**
Model-based offline RL algorithms which are able to handle image inputs are necessary for some environments.
**Describe the solut…
-
### Question
Hi, please state clearly in the documentation and dataset definition if in a time step "r_0" is consequence of "a_0"
With previous Offline RL libs, there has been some confusion wi…
-
What do we want out of our experiments? In the setting of offline RL, we want our algorithm to
1. Achieve reasonable success on the task
2. Show that adding distribution risk improves over vanilla …