I'm experiencing extremely slow training times while training BiFormer Small on a custom dataset with an image size of 224x244. The training takes up to a minute per batch with a batch size of 32 on an RTX 3070 GPU. I'm unsure if this is an expected training speed or if there might be an issue with my training setup implementation.
I have verified that the GPU is being utilized during training. Other models have not shown similar slowdowns on the same hardware and dataset.
Any guidance or suggestions to improve the training speed would be greatly appreciated!
I'm experiencing extremely slow training times while training BiFormer Small on a custom dataset with an image size of 224x244. The training takes up to a minute per batch with a batch size of 32 on an RTX 3070 GPU. I'm unsure if this is an expected training speed or if there might be an issue with my training setup implementation.
I have verified that the GPU is being utilized during training. Other models have not shown similar slowdowns on the same hardware and dataset.
Any guidance or suggestions to improve the training speed would be greatly appreciated!
Thank you!