Closed Ji4chenLi closed 1 year ago
Hi Jiachen,
Thanks for your interest in our project. To answer your questions:
We trained for 10 epochs in total.
The equation you wrote seems to be cyclical. We used a schedule that first linearly increases then monotonically decreases. A similar implementation can be found here.
We used "layer" to refer to transformer layer (block).
Yes, we computed success rate for each task averaged over 100 instances.
Our vectorized env implementation is based on this.
Yes, we used that to emulate workspaces that are free of robot arm occlusions.
Feel free to let me know if you have further questions.
Thank you so much for your response! My questions now get mostly addressed. The questions I left are:
min_lr
when performing the cosine annealing? Is it 1e-5 per here and Chinchilla?learning_rate
decreases to min_lr
? Is it step 17K or step 24K (7K + 17K)?Thank you so much for your response! My questions now get mostly addressed. The questions I left are:
- What's the
min_lr
when performing the cosine annealing? Is it 1e-5 per here and Chinchilla?- The Warmup Steps = 7K and the LR Cosine Annealing Steps = 17K as per Table 7. Could you let me know when the
learning_rate
decreases tomin_lr
? Is it step 17K or step 24K (7K + 17K)?
min_lr = 1e-7
.
Hi Yunfan,
Thank you so much for the great work! Since I'm trying to reproduce the results, I would like to ask some questions regarding the training and evaluation details.
task distribution
andlevel
, how many task instances did you sample? I assume the equation you used is $$SR = \frac{\text{number of success}}{\text{number of total task instances}} $$hide_arm_rgb
when making theenv
? Should we always set it toTrue
?Thanks and regards, Jiachen