Open przemb opened 9 months ago
I am also curious of reproducing the work to see if I can use it to its full potential. I can answer part of your question. It should be around 100k iterations since on page 23 in their paper they write: “According to our experiments, the training process of our ground truth encoder is easy to converge, and it usually takes only 1,000 iterations (stop training when the valid maxF is greater than 0.99). While the segmentation component of our model usually converges after around 100k iterations, and the whole training process takes less than 48 hours.”
For reproducing I also wonder, the initial values for the learning rate and optimizer were acording to the paper: "(initial learning rate lr=1e-3, betas=(0.9, 0.999), eps=1e-8, weight decay=0)". But it was not mentioned how or when the learning rate was changed. If I don't change it manualy the model stops converging around 15k iterations.
Hello, @xuebinqin, thanks for open sourcing your work. Could you please share some training tips and answer below questions?
I am using default configuration:
a) training from scratch: Let's suppose that I would like to train a model from scratch using DIS to get the same results as provided in
isnet.pth
.ite_num
needed to achieve the same results?b) fine-tuning: I would like to fine-tune a model on custom dataset from medical domain.
isnet.pth
orisnet-general
, which has higher performance, but was optimized)ISNetGTEncoder
?