I have tried multiple attempts over months to increase the depth of the nets layers (8- 10). If you watch the process seems fine with the individual training loss of each batch reducing. But at the end of the epoch it shows NAN i.e. infinity?
Although in one case I was able to "power through" the NaN, that hasn't happened since, despite trying up to150 epochs.
I am looking at this paper which seems to be addressing the problem.
I have tried multiple attempts over months to increase the depth of the nets layers (8- 10). If you watch the process seems fine with the individual training loss of each batch reducing. But at the end of the epoch it shows NAN i.e. infinity?
Although in one case I was able to "power through" the NaN, that hasn't happened since, despite trying up to150 epochs.
I am looking at this paper which seems to be addressing the problem.
http://torch.ch/blog/2016/02/04/resnets.html