pkuCactus / BDCN

The code for the CVPR2019 paper Bi-Directional Cascade Network for Perceptual Edge Detection
MIT License
341 stars 71 forks source link

用自己生成的数据训练loss高 #16

Closed tangsipeng closed 5 years ago

tangsipeng commented 5 years ago

感谢开源,我按自己的理解生成了一批数据用来训练,但是好像不太正常,希望得到答复。 数据:自己生成的数据,image数据是规则形状(不同比例长方形)的黑色边框混合高斯噪音背景,gt数据与image数据对应的是黑底白边的边框。分辨率不统一,大概1500*1000左右。 训练:所有参数都按照代码默认,希望可以训练得到边框的预测值,但是训练的时候loss非常高。 问题: 1、我这样生成数据是否正确? 2、不知道是这么高的loss否正常? 3、不知道有什么建议?

训练输出如下: 2019-07-01 09:13:10 - 19327 - train - 165: - iter: 20, lr: 1.000000e-06, loss: 4542657.586670, time using: 268.832860(13.441643s/iter) 2019-07-01 09:17:27 - 19327 - train - 165: - iter: 40, lr: 1.000000e-06, loss: 5015165.073022, time using: 257.739118(12.886956s/iter) 2019-07-01 09:21:52 - 19327 - train - 165: - iter: 60, lr: 1.000000e-06, loss: 5451473.621562, time using: 264.798099(13.239905s/iter) 2019-07-01 09:26:12 - 19327 - train - 165: - iter: 80, lr: 1.000000e-06, loss: 5684060.033437, time using: 259.854109(12.992705s/iter) 2019-07-01 09:30:38 - 19327 - train - 165: - iter: 100, lr: 1.000000e-06, loss: 5733884.890625, time using: 265.694983(13.284749s/iter) 2019-07-01 09:34:51 - 19327 - train - 165: - iter: 120, lr: 1.000000e-06, loss: 5772163.593750, time using: 252.914639(12.645732s/iter) 2019-07-01 09:39:14 - 19327 - train - 165: - iter: 140, lr: 1.000000e-06, loss: 5874098.355625, time using: 263.307687(13.165384s/iter) 2019-07-01 09:43:42 - 19327 - train - 165: - iter: 160, lr: 1.000000e-06, loss: 5776728.520313, time using: 267.739432(13.386972s/iter) 2019-07-01 09:47:55 - 19327 - train - 165: - iter: 180, lr: 1.000000e-06, loss: 5788587.094688, time using: 252.926788(12.646339s/iter) 2019-07-01 09:52:04 - 19327 - train - 165: - iter: 200, lr: 1.000000e-06, loss: 5926115.983125, time using: 249.590764(12.479538s/iter) 2019-07-01 09:56:26 - 19327 - train - 165: - iter: 220, lr: 1.000000e-06, loss: 5850768.974375, time using: 261.717345(13.085867s/iter) 2019-07-01 10:00:35 - 19327 - train - 165: - iter: 240, lr: 1.000000e-06, loss: 5650565.324375, time using: 249.568756(12.478438s/iter) 2019-07-01 10:04:53 - 19327 - train - 165: - iter: 260, lr: 1.000000e-06, loss: 5777168.478438, time using: 257.734875(12.886744s/iter) 2019-07-01 10:09:16 - 19327 - train - 165: - iter: 280, lr: 1.000000e-06, loss: 5784601.113750, time using: 262.375716(13.118786s/iter)

tangsipeng commented 5 years ago

补充:训练的时候加载了页面提供的vgg16的模型,没有加载发布的预训练模型

tangsipeng commented 5 years ago

看了论文以后做了一些修改,现在loss看起正常多了。 1、样本是白底黑边; 2、输入样本做了统一的resize 3、学习率保持在1e-6