Thanks for bring the work to public !! I have a question about experiments in Table 5.
In the paper, it is claimed that the training method of repVIT is identical to mobilenet-v3L, which consists of many modern training tricks. I believe the model used in Table 5 is also trained with these. It shows that resnet18 is a bit faster than repVIT-M1.1, but its down-stream task performance is much worse. Does the resnet18 model used here is also trained with mobilenet-v3L recipe, or it is only the original resnet18 model trained for 100 epoch without other tricks?
Thanks for your interest. The ResNet18 model in Table 5 is also trained under the same recipe as RepViT. We refer the results of ResNet18 in Table 5 from PVT and EfficientFormer, etc.
Hi,
Thanks for bring the work to public !! I have a question about experiments in Table 5.
In the paper, it is claimed that the training method of repVIT is identical to mobilenet-v3L, which consists of many modern training tricks. I believe the model used in Table 5 is also trained with these. It shows that resnet18 is a bit faster than repVIT-M1.1, but its down-stream task performance is much worse. Does the resnet18 model used here is also trained with mobilenet-v3L recipe, or it is only the original resnet18 model trained for 100 epoch without other tricks?