Walter0807 / MotionBERT

[ICCV 2023] PyTorch Implementation of "MotionBERT: A Unified Perspective on Learning Human Motion Representations"
Apache License 2.0
1.01k stars 123 forks source link

Some questions about training #72

Closed liuxing007 closed 1 year ago

liuxing007 commented 1 year ago

Great work! I have a few questions:

(1) In the second version, the MB_train_h36m.yaml file was modified to set dim_feat as 512, whereas the paper mentions it as 256. Will this change improve the results?

(2) Without making any modifications, I ran the latest code for training 3D human pose estimation from scratch. However, the model file size in the checkpoint is 486.2MB, which is different from the 162MB model file uploaded by your team. Nevertheless, I am able to evaluate your model file and our trained 486.2MB model file directly. I'm curious to know the reason behind this.

(3) How many epochs did you train the model for? I have been unable to achieve the results mentioned in the paper. The best reported MPJPE is 39.4mm.

Walter0807 commented 1 year ago
  1. It is the main difference between MB and MB_lite. The paper version is also 512 (updated).
  2. The saved checkpoint also contains training parameters (gradients, optimizer states, etc.), which are not needed for inference, and therefore we removed them in the release.
  3. The number of training epochs is in the config. Did you use the same batch size and learning rate?
liuxing007 commented 1 year ago

QQ截图20230720105757 This is the result of training with dim_feat=256. The best reported MPJPE is 39.9mm. batch size and learning rate are the same.

liuxing007 commented 1 year ago

图片 This is the result of training with dim_feat=512. The best reported MPJPE is 39.4mm. batch size and learning rate are the same.