Open shenxiaojian opened 1 month ago
Thank you for your interest in my work and for your detailed question.
You are correct in noting that the upgraded PopDanceSet has significantly increased in size, now comparable to the AIST++ dataset. After applying data augmentation, the actual amount of training data indeed exceeds that of the AIST++ dataset, which contributes to the longer training time you are experiencing.
I have experimented with many hyperparameters during my work, and there were no substantial changes. Therefore, you can continue using the same hyperparameters as mentioned in the paper (i.e., a learning rate of 0.001, betas of 0.02, 0.08, and 0.01, with an epsilon of 1e-8) to reproduce the results.
I hope this helps clarify your concerns.
Hi,
Thanks for your excellent work! I've been working on reproducing your paper using two A800 GPUs, but I've encountered some issues. In the PopDanceSet_Description, it's mentioned that the PopDanceSet has been upgraded. With
repeat_count=1
, I found that one epoch takes about 5 minutes to run, meaning 2000 epochs would take approximately 166 hours. This significantly differs from the 66 hours mentioned in the paper.I suspect that this might be due to substantial differences between the updated dataset and the previous one.
Could you please clarify:
Here is the table from the PopDanceSet_Description:
Thank you very much for your assistance.