Open ZhihuaGao opened 7 years ago
Usually if you encounter loss going up rapidly in the beginning, you should lower the learning rate initially and then gradually make the learning rate higher. People usually call it "warm up".
Thank for your advise,I will have a try.
I see "value:20", Is it appropriate? layer { name: "conv3_3_norm" type: "Normalize" bottom: "conv3_3" top: "conv3_3_norm" norm_param { across_spatial: false scale_filler { type: "constant" value: 20 } channel_shared: false } }
@weiliu89 @NickMinYang Yeah,that is a question,I really do not sure about the value about Normalization layer. Theoretical should I calculate the feature scale of conv3_3,conv4_3,fc6,conv6_2,conv7_2,conv8_2,conv9_2 mentioned in the parsenet and then infer the value?
@AresGao can you share the code of the ssd_pascal.py added the con3_3 layer?
Excuse me, is your method feasible at last?
In order to detect small targets,I add new mbox_ layers after conv3_3; The configuration is
But when I start training I find that the leaning rate should decay to 0.0001 or I will get a nan;And finally the map reduced by 7 points;Anyone can help me?