Open Grassyue opened 5 years ago
Each batch will first split into N (depending on the value of split_batch
) sub-batches before feeding to the network. The larger split_batch
will lead to less GPU memory usage during training.
Thanks for your reply,but why not directly set the batch size to a small number? what is the difference?
When you set a large batchsize, the option split_batch
make it possible to update the parameters only once.
When I use the option split_batch
, the accuracy was much worse on the test set, what is the reason for this?
Hi, thank you for your sharing and sorry to bother you, i am a new guy in super-resolution, can you explain the meaning of "split_batch" in the code? thank you sooo much!