cremebrule / digital-cousins

Codebase for Automated Creation of Digital Cousins for Robust Policy Learning
https://digital-cousins.github.io
Apache License 2.0
122 stars 12 forks source link

reproduce the released checkpoints #13

Open andyaloha opened 2 weeks ago

andyaloha commented 2 weeks ago

Could you please provide the process for reproducing the training of the 'cousin_ckpt.pth' and 'twin_ckpt.pth' files? Thank you.

cremebrule commented 2 weeks ago

Hi @andyaloha ,

Thanks for reaching out! We trained our policies on either exclusively the digital twin (twin_ckpt.pth) or 8 cousin models (cousin_ckpt.pth), collecting a total of 10000 demos in either case. We then trained both models using BC-RNN and deployed them immediately on different evaluation cases (i.e.: digital twin, or other unseen cabinets).

You can reproduce the hyperparameters used by inspecting the environment config file in the checkpoint itself!

RogerDAI1217 commented 1 week ago

Hi @andyaloha, to save your time, here is the models we used for the twin v.s. cousins experiments on door opening task.

Use DINO to select digital cousins:

Use DINO + GPT to select digital cousins:

We trained BC-RNN with different hyperparameters. The average success rates are reported in Table 3, Table 4, Table 5, Figure 7, Figure 13, and Figure 14 in Appendix.

andyaloha commented 1 week ago

@cremebrule @RogerDAI1217 Thanks. I have trained a valid policy according to your instruction with twin of kdbgpm and 8 cousins in the config of twin_ckpt.pth. While in evaluation, it seems that the target object must be one of those in the loaded scene? This scene is specified in the demos collection step by scene_path, which is saved into the .hdf5 file and loaded as the test scene in the evaluation step. How can I change the target object to others not in the loaded scene (such as different hold-out cousins)?

cremebrule commented 1 week ago

Hi @andyaloha ,

Great question. I believe we already have this implemented -- you can simply set the eval_category_model_link_name in the evaluation script examples/4_evaluate_policy.py. Can you try that?

andyaloha commented 6 days ago

@cremebrule I have tried it. No matter how I change the target object by eval_category_model_link_name, the final target object is always the same kdbgpm. It seems that the target object is fixed by the scene_path as in the demos collection step, which is saved into the .hdf5 file and loaded as the test scene in the evaluation step.