Closed idonotlikemondays closed 2 days ago
Thanks for the issue. I see some of the datasets are outdated and need to be regenerated due to a small change in the environment.
I'll also investigate the low conversion success rates
Thanks for the issue. I see some of the datasets are outdated and need to be regenerated due to a small change in the environment.
I'll also investigate the low conversion success rates
Thanks a lot! I really appreciate your help! I also want to ask if it's possible to add demonstrations for more tasks? e.g pushT or TwoRobot related tasks?
Yeah we plan to. Some people will work on writing motion planning solutions / generating teleop demos for more tasks over time.
PushCube replay works as intended.
StackCube demos are re-uploaded now.
What is the script you using to train your policy? It does seems strange joint space control does worse but pd ee delta pos should work fine as well
What is the script you using to train your policy? It does seems strange joint space control does worse but pd ee delta pos should work fine as well
I use diffusion policy script to train the policy. And pd_joint_delta_pos
control mode works fine but basically it is not available with pd_ee_delta_pos
control mode.
I noticed when I replay PushCube task with the pd_ee_delta_pos
control mode, the conversion success rate is not good (compare to the pd_joint_delta_pos
control mode, which has nearly 100% conversion success rate), I guess it might be the reason?
PushCube replay works as intended.
StackCube demos are re-uploaded now.
Hi,
I did some experiments with the new version of StackCube demos (both pd_ee_delta_pos
and pd_joint_delta_pos
), but it shows poor performance.
I want to ask whether this is due to the demo itself or because the algorithm (diffusion policy in my case) is not sufficient to support this task? However, diffusion policy performs well in Peg Insertion... 🤔
Thanks a lot!
can you try increasing the max episode steps? the demos are kind of slow so imitating them results in a policy that takes rather long to get success.
For example stack cube max episode steps is 50 (tuned for RL) but the motion planning demos might average around 100 steps
can you try increasing the max episode steps? the demos are kind of slow so imitating them results in a policy that takes rather long to get success.
For example stack cube max episode steps is 50 (tuned for RL) but the motion planning demos might average around 100 steps
Yes, these results are based on max_episode_steps=100
.
I run it with config:
total_iters: 30000
batch_size: 128 # (since 1024 is kinda slow)
max_episode_steps: 100
num_demos: 100
num_diffusion_iters: 5 # (in original code is 100, but I try 5 in other tasks and which get better performance and faster running speed)
I would try max_episode_steps=200
latter, and do you have any other idea to deal with the issue?
Actually it turns out it should be more than 100 max episode steps. Looking at the eval videos each the model does pretty well but is just slow because the demos are a bit slow. with max of 200 steps it works better
It seems I need to establish a solid set of recommended max episode steps. Especially given that most pure offline imitation learning algorithms are not meant to optimize for fast solving and instead to mimic demo distribution as much as possible, I may just recommend users to set max episode steps to about 1.5x the mean episode steps of the demonstration data and then only check the success_once evaluation metric.
Also for push cube I find no issue with training on that task. Using the example.sh script in the diffusion policy baseline folder works for me.
Hi,
1. Unsuccessfully replay trajectory in task "StackCube_v1" I tried to replay the trajectory of PushCube-v1 but failed.
I tried
pip uninstall pytorch_kinematics_ms
and reinstalled and upgraded to the latest version of ManiSkill (3.0.0b10), but it still doesn't work. I would like to ask if you have any suggestions to solve this problem?2. Low convert rate and worse performance in
pd_ee_delta_pos
compare topd_joint_delta_pos
By the way, among the five tasks ['PickCube-v1', 'PushCube-v1', 'StackCube-v1', 'PegInsertionSide-v1', 'PlugCharger-v1'], withpd_ee_delta_pos
as the control mode, except for PickCube which has a 100% success rate, PushCube also has a relatively high failure rate. Additionally, the results of running diffusion policy at the end are not ideal (as shown in the figure). The remaining three tasks all display "not replayed successfully." Do you have any effective solutions to address this issue? (The former ispd_joint_delta_pos
, and the latter ispd_ee_delta_pos
. Theoretically, the latter should be easier to learn than the former?)Thanks a lot for your help!!
Best, Zhenyu