-
Hi, thanks for sharing your great work. However, I am confused about the rollout generation process.
As I see in the code, the agent can access to a pre-defined terminal function to cut down the un…
-
To look at locations
```sql
SELECT detector_name, street1, street2, bluetooth_id, wifi_id, detector_name_old, project, date_start, date_end, latitude,
longitude, "PX", index1, loc
FROM blueto…
-
Hi @justinjfu @aviralkumar2907, I want to train Mujoco-Gym Continuous Control Tasks with Image Observations.
I figured out that `image_envs` branch support Image Observations. But when I did `gym.ma…
-
Imagine that we use ReAgent to train a personalization policy, and the workflow is as follows:
1. We collect a number of user interaction histories (episodes) and train a DQN model in offline (Batch…
-
Sorry, I have two new problems through learning your program.
1. How do you define offline and online.
2. Why does (offline training online -------- training training -------- converting data) nee…
-
link to in contribution guidelines, what is in scope etc
-
Dear author, I am reproducing this code recently. I have some questions to ask you:
1. Why is the model training "MAX_EPISODE_NUM" set to 8000 ? Is the result better the more times you train? Will i…
-
![image](https://github.com/Zhendong-Wang/Diffusion-Policies-for-Offline-RL/assets/87383739/771568ad-84af-4db4-8e21-c5c2fda8701c)
the above is the description of the effect of timesteps of diffusion …
-
-
I am interested in training these models on a completely different dataset (i.e. not using DM Control or MuJoCo). I have recorded simulation data from Isaac Sim and I would like to train the models in…