Closed aymuos15 closed 1 month ago
- For nnUNet, did you use the resnet version?
- Why use Cross-Entropy as the loss? -- This I suppose adversely affects most of the networks.
- Are there any ablations based on the loss?
- Why are there no overlap metrics reported?
- I also think testing on the BraTS-Mets dataset would be a nice test.
Also, Minor typo: 'of training except that nnUNet utilized official settings [13] for training.' Should be: 'of training except that of nnUNet utilized official settings [13] for training.'
Thanks!
Thank you for your interest on our work. I will address your questions one by one:
If I was incorrect about the "overlap metrics" or other points, could you please provide more information?
2/3. I would argue saying Dice + CE is the most common one. I do agree about not being strict about the loss, but generally segmentation tasks do well with Dice + CE. Obviously it may show an increase for your model as well, but at least to me, Dice + CE would be a fairer evaluation. Just to go on your point, efficiency isn't a discussed objective in your paper either right?
Apologies, I meant to say boundary metrics (Like HD95).
Sounds good!
Really appreciate the time taken to answer the questions!
- I was talking about this -- ResNet Presets
2/3. I would argue saying Dice + CE is the most common one. I do agree about not being strict about the loss, but generally segmentation tasks do well with Dice + CE. Obviously it may show an increase for your model as well, but at least to me, Dice + CE would be a fairer evaluation. Just to go on your point, efficiency isn't a discussed objective in your paper either right?
- Apologies, I meant to say boundary metrics (Like HD95).
- Sounds good!
Really appreciate the time taken to answer the questions!
Thanks for your reply. We used nnU-Net (org.) from thier repo. For the loss function, it's worth trying Dice + CE during training. We plan to implement more metrics in the future work. We noticed that several online implemetations of metrics (e.g., HD95) don't support vectors with batch dimensions well, leading to slow validation if performed every epoch.
I think you should use the resnet presets. They offer better scores. See this discussion about a different paper on some of the few challenge datasets -- https://openreview.net/forum?id=qmN9v3O69J
And yes, that's fair. I was only talking about test metrics.
Thank you for answering all the questions and great work! Really interesting read:)
Also, Minor typo: 'of training except that nnUNet utilized official settings [13] for training.' Should be: 'of training except that of nnUNet utilized official settings [13] for training.'
Thanks!