Closed WNjun closed 6 days ago
Hi @WNjun , thanks for your attention to our work!
TinyViT-5M is a small model trained with a drop path rate of 0. Therefore, the argument DROP_PATH_RATE should be set to 0.0
Besides, TRAIN.LAYER_LR_DECAY could be set to 0.8
Hi @wkcn, thank you for your prompt reply!
I've added DROP_PATH_RATE = 0
and LAYER_LR_DECAY = 0.8
to the config file. The results improved slightly from 94.2% to 94.73%, but they still don't surpass the 224 model's accuracy of 94.89%. Do you have any further suggestions? Also, my dataset has 11 classes; could this be affecting the results?
@WNjun The head's weight corresponding to the 11 classes can be inherited from the 1k classifier head. You can refer to the implementation: https://github.com/wkcn/TinyViT/blob/main/utils.py#L75
Besides, the problem of overfitting could be avoided by decreasing LAYER_LR_DECAY
.
Thank you for your help! I'll definitely look into these and try them out later. Thanks again!
Hi,
I've recently trained a custom TinyViT-5M-22kto1k model and attempted to finetune it from an image resolution of 224 to 384. However, I observed no improvement in the model's performance post finetuning. I'm wondering if this could be due to the limitations of the model's size, or perhaps I might be missing something in my configuration.
Here is the YAML configuration for the model:
Could you please help me identify any potential issues or suggest improvements to enhance the finetuning process?
Thank you for your assistance!