Closed wnzhyee closed 6 years ago
This might not be an 'out of memory' issue. We've seen this occasionally also in our experiment, and we don't know the reason.
You can simply kill the program and rerun it.
@wnzhyee i have only one 12GB gpu too.what's your solution?could you please tell me the method?:)thanks
I see that in train.sh, it use 4x12G GPUs for training, but i only have one 12G GPU(1080 Ti), so how can i reduce the GPU cost?
i try change '-nGPU 4' to '-nGPU 1', and add '-trainBatch 4' in train.sh, but it seems doesn't work because the model 'converting the model to CUDA' again and again, just like:
So how can i do?Thanks.