Open faybak opened 4 years ago
there are a lot of nans values on my training. Is it normal ? I want to train 36 classes with a dataset containing very small objects on images with different size. i get the anchor using what it said in this repo. Can you help me please
here is my cfg file: [net]
batch=64 subdivisions=64 width=1088 height=1088 channels=3 momentum=0.9 decay=0.0005 angle=0 saturation = 1.2 exposure = 1.2 hue=.07
learning_rate=0.001 burn_in=1000 max_batches = 72000 policy=steps steps=57600,64800 scales=.1,.1 letter_box=1
[convolutional] batch_normalize=1 filters=32 size=3 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=64 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=32 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=64 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=128 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=64 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=128 size=3 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=1024 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=1024 size=3 stride=1 pad=1 activation=leaky
######################
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=1024 activation=leaky
[convolutional] size=1 stride=1 pad=1 filters=123 activation=linear
[yolo] mask = 12,13,14 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
[route] layers = -4
[upsample] stride=2
[route] layers = -1, 61
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=512 activation=leaky
[yolo] mask = 9,10,11 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
[route] layers = -1, 36
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=256 activation=leaky
[yolo] mask = 6,7,8 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
###############
[route] layers = -1, 11
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=128 activation=leaky
[yolo] mask = 3,4,5 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
[route] layers = -1, 4
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=64 activation=leaky
[yolo] mask = 0,1,2 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
It's normal because avgloss isn't 0.
@lsd1994 can you please let me know how you calculated avgloss are not 0 by looking at that screenshot ?
there are a lot of nans values on my training. Is it normal ? I want to train 36 classes with a dataset containing very small objects on images with different size. i get the anchor using what it said in this repo. Can you help me please
here is my cfg file: [net]
Testing
batch=1
subdivisions=1
Training
batch=64 subdivisions=64 width=1088 height=1088 channels=3 momentum=0.9 decay=0.0005 angle=0 saturation = 1.2 exposure = 1.2 hue=.07
learning_rate=0.001 burn_in=1000 max_batches = 72000 policy=steps steps=57600,64800 scales=.1,.1 letter_box=1
[convolutional] batch_normalize=1 filters=32 size=3 stride=1 pad=1 activation=leaky
Downsample
[convolutional] batch_normalize=1 filters=64 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=32 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=64 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
Downsample
[convolutional] batch_normalize=1 filters=128 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=64 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=128 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=64 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=128 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
Downsample
[convolutional] batch_normalize=1 filters=256 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
Downsample
[convolutional] batch_normalize=1 filters=512 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
Downsample
[convolutional] batch_normalize=1 filters=1024 size=3 stride=2 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=1024 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=512 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=1024 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=512 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=1024 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
[convolutional] batch_normalize=1 filters=512 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 filters=1024 size=3 stride=1 pad=1 activation=leaky
[shortcut] from=-3 activation=linear
######################
[convolutional] batch_normalize=1 filters=512 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=1024 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=1024 activation=leaky
[convolutional] batch_normalize=1 filters=512 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=1024 activation=leaky
[convolutional] size=1 stride=1 pad=1 filters=123 activation=linear
[yolo] mask = 12,13,14 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
[route] layers = -4
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[upsample] stride=2
[route] layers = -1, 61
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=512 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=512 activation=leaky
[convolutional] batch_normalize=1 filters=256 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=512 activation=leaky
[convolutional] size=1 stride=1 pad=1 filters=123 activation=linear
[yolo] mask = 9,10,11 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
[route] layers = -4
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[upsample] stride=2
[route] layers = -1, 36
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=256 activation=leaky
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=256 activation=leaky
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=256 activation=leaky
[convolutional] size=1 stride=1 pad=1 filters=123 activation=linear
[yolo] mask = 6,7,8 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
###############
[route] layers = -4
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[upsample] stride=2
[route] layers = -1, 11
[convolutional] batch_normalize=1 filters=64 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=128 activation=leaky
[convolutional] batch_normalize=1 filters=64 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=128 activation=leaky
[convolutional] batch_normalize=1 filters=64 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=128 activation=leaky
[convolutional] size=1 stride=1 pad=1 filters=123 activation=linear
[yolo] mask = 3,4,5 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0
[route] layers = -4
[convolutional] batch_normalize=1 filters=128 size=1 stride=1 pad=1 activation=leaky
[upsample] stride=2
[route] layers = -1, 4
[convolutional] batch_normalize=1 filters=32 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=64 activation=leaky
[convolutional] batch_normalize=1 filters=32 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=64 activation=leaky
[convolutional] batch_normalize=1 filters=32 size=1 stride=1 pad=1 activation=leaky
[convolutional] batch_normalize=1 size=3 stride=1 pad=1 filters=64 activation=leaky
[convolutional] size=1 stride=1 pad=1 filters=123 activation=linear
[yolo] mask = 0,1,2 anchors = 7,9, 10,19, 11,29, 27,16, 14,32, 12,40, 17,40, 20,53, 27,71, 44,145, 138,56, 60,265, 207,95, 101,581, 115,835 classes=36 num=15 jitter=.3 ignore_thresh = .7 truth_thresh = 1 random=0