Closed ManUtdMoon closed 8 months ago
Hi, in our latest exps we found that simply training the models separately would yield similar/even better performance. Joint training is not needed any more.
Hi, in our latest exps we found that simply training the models separately would yield similar/even better performance. Joint training is not needed any more.
Hi Zhou, thank you for your quick reply!
I think seperate training needs relabeling of the keypose. Could you please show me how it is done in the code? Thank you!
What do you mean relabeling? We use the same strategy for both act3d and chained diffuser for extracting keyframes.
On Tue, Dec 26, 2023 at 4:46 PM Dongjie Yu @.***> wrote:
Hi, in our latest exps we found that simply training the models separately would yield similar/even better performance. Joint training is not needed any more.
Hi Zhou, thank you for your quick reply!
I think seperate training needs relabeling of the keypose. Could you please show me how it is done in the code? Thank you!
— Reply to this email directly, view it on GitHub https://github.com/zhouxian/act3d-chained-diffuser/issues/11#issuecomment-1869846484, or unsubscribe https://github.com/notifications/unsubscribe-auth/AEV4V6OHGFX73HADGXKNZHTYLNVWZAVCNFSM6AAAAABBDMKP3GVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQNRZHA2DMNBYGQ . You are receiving this because you commented.Message ID: @.***>
The paper mentioned that goal gripper pose is not ground truth but predicted by the action detector. Therefore, during training, I think the target keypose should be relabeled by act3d instead of taking from the ground truth. Is there something wrong with my understanding?
You can simply train them separately and use gt pose during training. At inference time act3d can be used to key pose prediction.
On Tue, Dec 26, 2023 at 5:17 PM Dongjie Yu @.***> wrote:
The paper mentioned that goal gripper pose is not ground truth but predicted by the action detector. Therefore, during training, I think the target keypose should be relabeled by act3d instead of taking from the ground truth. Is there something wrong with my understanding?
— Reply to this email directly, view it on GitHub https://github.com/zhouxian/act3d-chained-diffuser/issues/11#issuecomment-1869857650, or unsubscribe https://github.com/notifications/unsubscribe-auth/AEV4V6OBMFLKE7FVFYSXMYDYLNZKTAVCNFSM6AAAAABBDMKP3GVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQNRZHA2TONRVGA . You are receiving this because you commented.Message ID: @.***>
Thank you for your answers! Have a nice day!
Dear authors,
Thank you for your inspiring work!
I noticed that in the ChainedDiffuser paper your mentioned that "...train both the action detector and the trajectory diffuser jointly" and "we train the first 2 terms till convergence, and then add the 3rd term for joint optimization". However, I did not see that there are codes for joint optimization because the only model in
main_trajectory.py
is aDiffusionPlanner
.Would you please explain more about the actual joint training of Act3d and DiffusionPlanner?
Regards, Dongjie