open-mmlab / mmpose

OpenMMLab Pose Estimation Toolbox and Benchmark.
https://mmpose.readthedocs.io/en/latest/
Apache License 2.0
5.55k stars 1.21k forks source link

All losses are Zero. How should we USE finetuning? #1806

Open Msameim181 opened 1 year ago

Msameim181 commented 1 year ago

I'm using a hand dataset that is different from other hand images, so the MMDetection would not detect that. So I train it with YOLOX and I got high accuracy. And then I want to use hand Keypoint to detect keypoint on hand images. I use Label-studio and fixed annotation to coco-format (or should we better say 'FreiHandDataset'). The Training Starts but no improvement at all.

I use this res50_freihand2d_224x224.py config and I load this Model [res50_freihand_224x224-ff0799bc_20200914.pth]() for pre-train model.

The first question is how to use the YOLOX I trained in this training? Should I use it? Second, Why is this going nowhere?

image

image

image

image

ly015 commented 1 year ago

Thanks for using mmpose. Could you please share the full configs used in the experiment? And could you please check if the bounding boxes obtained by either Label Studio or YOLOX have been added to the annotation files for the keypoint detection task? It seems that you are using a top-down keypoint model so the bbox information is necessary.

Msameim181 commented 1 year ago

All modified config was shown in images. I used this this res50_freihand2d_224x224.py config for base.

All Data that is used have bbox and 21 keypoints for hand. I use the same JSON file for training the YOLOX.

Msameim181 commented 1 year ago

Thanks for using mmpose. Could you please share the full configs used in the experiment? And could you please check if the bounding boxes obtained by either Label Studio or YOLOX have been added to the annotation files for the keypoint detection task? It seems that you are using a top-down keypoint model so the bbox information is necessary.

Q: The coordination of keypoints should be in pixel or percentage or something else?

jin-s13 commented 1 year ago

The coordination of keypoints should be in pixel.

Normally we will use the ground-truth bounding box for training. And during inference, we can use YOLOX detection results.

Msameim181 commented 1 year ago

Thanks for the update. I Changed my dataset format to COCO and use this format as DataClass in the config. Now I'm using HRNet and everything works fine. Thanks.

Msameim181 commented 1 year ago

And for The Final Question, How could we Contribute to the Pre-train models and datasets?