XunshanMan / MVGFormer

This is the official implementation of the work presented at CVPR 2024, titled Multiple View Geometry Transformers for 3D Human Pose Estimation (MVGFormer).
Apache License 2.0
29 stars 0 forks source link

Regarding Results after Epoch 1 Training #4

Open aviralchharia opened 1 month ago

aviralchharia commented 1 month ago

Hi, thanks for the great work! I tried training the model and got two tables as the result after the first epoch.

  1. The two tables show different values of $AP{25}$, $AP{50}$, $AP{75}$, $AP{100}$, etc (Same for Recall values). Also two different values of $MPJPE$ are printed. I assume the paper reports the first table. What does the second table here represent?
  2. Moreover, wandB only shows the Table 2 values. Am I missing something?

image

aviralchharia commented 1 month ago

image

XunshanMan commented 1 month ago

Hi, thanks for the question! The difference between the two tables is whether to use NMS before outputting the results. NMS is necessary (see Table 1) because we start with 1024 queries (pose assumptions), and even after filtering by classification scores, multiple queries remain linked to the same person. After NMS, the redundant queries can be effectively filtered. For Wandb, you can add the metric in Table 1 into it for plot supervision.

SWWdz commented 3 weeks ago

Hi, thanks for the great work! I tried training the model and got two tables as the result after the first epoch.

  1. Moreover, wandB only shows the Table 2 values. Am I missing something?

image

Can I ask you how to train this model? What's your cuda version, gpu and your environment?

XunshanMan commented 2 weeks ago

Hi, thanks for the great work! I tried training the model and got two tables as the result after the first epoch.

  1. Moreover, wandB only shows the Table 2 values. Am I missing something? image

Can I ask you how to train this model? What's your cuda version, gpu and your environment?

Please refer to the README Training part for the details: https://github.com/XunshanMan/MVGFormer?tab=readme-ov-file#2-training.

Here is a runnable environment for reference: