Open jeasung-pf opened 4 years ago
That is kind of right. I added these loss_weights in order to encourage the network to correct errors in the third word, since the dataset mainly consists of images with three words. This was just a performance measure and everything should work fine with loss equally weighted loss weights.
Hello.
I am now training the FSNS datasets with images containing 5 or more boxes and the thing is that when calculating losses in your model, there are preconfigured weights to be multiplied with the loss on each bounding boxes and that makes an index out of bound error when calculating the loss on the fifth bounding box. Below are code blocks you wrote.
Does it mean that usually the size and complexity of decoding sequences in the third box is bigger than any other boxes in the data? If so, did you conduct experiments with equally weighted bounding boxes?