Vegetebird / StridedTransformer-Pose3D

[TMM 2022] Exploiting Temporal Contexts with Strided Transformer for 3D Human Pose Estimation
MIT License
335 stars 37 forks source link

how to use 27 frames #17

Closed parkjin0903 closed 2 years ago

parkjin0903 commented 2 years ago

Thank you for providing the inference code. I am very interested in your paper. According to the paper and code, it seems like using 351 frames apparently. I tried to modify vis.py code to use 27 frames. Unfortunately I faced a hardship. The following is the error code that I got. Could you give me a instruction?


size mismatch for Transformer.pos_embedding: copying a param with shape torch.Size([1, 351, 256]) from checkpoint, the shape in current model is torch.Size([1, 27, 256]). size mismatch for Transformer_reduce.model.pos_embedding_1: copying a param with shape torch.Size([1, 351, 256]) from checkpoint, the shape in current model is torch.Size([1, 27, 256]). size mismatch for Transformer_reduce.model.pos_embedding_2: copying a param with shape torch.Size([1, 351, 256]) from checkpoint, the shape in current model is torch.Size([1, 27, 256]). size mismatch for Transformer_reduce.model.pos_embedding_3: copying a param with shape torch.Size([1, 351, 256]) from checkpoint, the shape in current model is torch.Size([1, 27, 256]).

SonNguyen2510 commented 2 years ago

You need to retrain the model using 27 frames yourself first then using that model for your purpose.

Vegetebird commented 2 years ago

@SonNguyen2510 is right

parkjin0903 commented 1 year ago

Thank you !