Open aman-cc opened 1 year ago
I'm training the BLIP captioning model on a custom dataset with 80:20 split (train:val). The training time for the epoch is ~6hrs while the val step takes ~42 hrs. Any idea why this is the case? Batch size seems to be the same for both steps.
I'm training the BLIP captioning model on a custom dataset with 80:20 split (train:val). The training time for the epoch is ~6hrs while the val step takes ~42 hrs. Any idea why this is the case? Batch size seems to be the same for both steps.