Closed noisewm closed 5 years ago
hello, it is probably not enough information to reproduce your problem... Can you try to run the following code which is tested, https://circleci.com/gh/Borda/keras-yolo3/265 ?
Thx, training works just fine with VOC2007.
Seems that problem is in my dataset, will try to implement custom callback and look into the whole batch that results in NaN loss.
You do not have custom callback, you can just run debugger and place break-point at the end of generation an augmented image with bounding box...
@noisewm I believe that it is solved for now, but if you have any further question, feel free to reopen this issue...
Hello,
I'm trying to migrate from qqwweeee's yolo implementation, but training with my own dataset results in loss: NaN (in qqwweee i can train just fine with same dataset). Already implemented basic null checks, and also my dataset goes through automated checks (so every image present, every box checked by its coordinates, class present in classes file) before training.
Already tried with different batch sizes -- same effect. Time for loss function to become NaN seems random (almost always in 1st epoch).
I think that problem somehow related to data generator... Maybe you can suggest a proper way to debug it?