Open Erichen911 opened 3 years ago
@Erichen911 1200 is not enough! Off by 3 orders of magnitude at least!
@Erichen911 I would recommend getting a huge amount of images, preferrably a million at least, and then doing self-supervised learning with BYOL, before training on your tiny training set
otherwise, just use Ross' pretrained model!
@Erichen911 can you share your code?> tks
Does anyone train Swin transformers with different image sizes? Like my image size is 112x112, it never works on this size
@lucidrains @Erichen911 can you share the train.py which you guys r using for custom data or any reference ??
Hey guys. First of all. This is a great job and thanks to the authors. Then my question is... Recently I use this code on my own dataset. A simple binary-classification problem. The performance on the training dataset is good, but not as well as the validation dataset. The Loss curve is...
My model is model = ViT( dim=128, image_size=224, patch_size=32, num_classes=2, depth=12, heads=8, mlp_dim=512, channels=3, )
Training Dataset has 1200+ images, Validation Dataset has 300+ images.
Can someone give me some suggestions, how to solve this problem?
I think there are several possibilities. Maybe I need a pretrained model? Or I did the wrong way in the training of transformer model?