Closed iamrishab closed 4 years ago
We use 4 x Tesla P40 while training, maybe you can reduce the batch size.
Yes, I was changing a different batch (batch_size_v
) which is not used in the dataloader.
Changing batch_size_t
actually worked. Thanks!
We use 4 x Tesla P40 while training, maybe you can reduce the batch size.
Can you please share the details system configuration for model training using default hyperparemeters in the config.py
? tia
Never-mind. I figured it out. Thanks!
Hi! Can you please tell me what hardware configuration are you using for training? Especially for Step 4 and Step 5:
Currently, I am using 8 K80s but still getting
CUDA out of memory
while training. Thanks!