Closed mohsij closed 6 years ago
The best you can do is set
>> opts.train.numSubBatches=5
(equal to opts.train.batchSize).
That will make training a little slower but use much less memory.
I hope this solves your issue, with 8GB and batches split in sub-batches it should fit in the memory. You can also try smaller splits, ie numSubBatches=2 or 3.
I went through the second paper and it says the training was done on a TITAN X with 12GB of memory. I tried running the training code on a system with GTX 1070 with 8GB of memory and get the following error even if I reset the GPU memory at the start:
Is there a way to solve this by using a smaller subset of the training data or some other way? Thanks!