MIV-XJTU / ARTrack

Apache License 2.0
228 stars 33 forks source link

I wanna reproduce ARTrack-L-384 #31

Closed XJLeoYu closed 3 months ago

XJLeoYu commented 10 months ago

but seems that the files you provided didn't match,it report bug: dim 1024 should be divided by num_heads 12

AlexDotHam commented 10 months ago

Sorry, the num_heads should be 16, i will change that soon.

XJLeoYu commented 10 months ago

When I changed the num_heads back to 16 in config.py,it's still bugy

AlexDotHam commented 10 months ago

微信图片_20231114205445 微信图片_20231114205450 I have updated the YAML, on my server, and it is executing normally (it should be noticed that I only have 3090 for training, so I reset the batch_size to 2). Maybe there is something wrong with your operation. If you have any other questions, you can mail me and add my WeChat.

XJLeoYu commented 10 months ago

I was just planing to inference not traning,not pretty sure how to find you mail or wechat account on github

AlexDotHam commented 10 months ago

You can mail me yfbai@stu.xjtu.edu.cn

Jasper0420 commented 3 months ago

您好,我在使用384给出的权重推理时出现以下shape不匹配的问题?

size mismatch for pix_head.decoder.layers.3.multihead_attn.in_proj_bias: copying a param with shape torch.Size([3072]) from checkpoint, the shape in current model is torch.Size([2304]).
        size mismatch for pix_head.decoder.layers.3.multihead_attn.out_proj.weight: copying a param with shape torch.Size([1024, 1024]) from checkpoint, the shape in current model is torch.Size([768, 768]).
        size mismatch for pix_head.decoder.layers.3.multihead_attn.out_proj.bias: copying a param with shape torch.Size([1024]) from checkpoint, the shape in current model is torch.Size([768]).
        size mismatch for pix_head.decoder.layers.3.norm_3.weight: copying a param with shape torch.Size([1024]) from checkpoint, the shape in current model is torch.Size([768]).
        size mismatch for pix_head.decoder.layers.3.norm_3.bias: copying a param with shape torch.Size([1024]) from checkpoint, the shape in current model is torch.Size([768]).
        size mismatch for pix_head.decoder.layers.3.mlpz.fc1.weight: copying a param with shape torch.Size([4096, 1024]) from checkpoint, the shape in current model is torch.Size([3072, 768]).
        size mismatch for pix_head.decoder.layers.3.mlpz.fc1.bias: copying a param with shape torch.Size([4096]) from checkpoint, the shape in current model is torch.Size([3072]).
        size mismatch for pix_head.decoder.layers.3.mlpz.fc2.weight: copying a param with shape torch.Size([1024, 4096]) from checkpoint, the shape in current model is torch.Size([768, 3072]).

请问应该如何修改yaml文件或者其他配置吗?

AlexDotHam commented 3 months ago

It seems like you use the ARTrack-L-384 model checkpoints with a ARTrack-B-384 yaml, you can check the yaml you use to match the checkpoints.

Jasper0420 commented 3 months ago

Thanks! Good luck!