Early stopping focused on the validation loss metric would be nice. From my minimal understanding, it would be hard to do with the current system of extending BaseMetric. At least with IoUMetric, there is no ability to early stop based on the validation loss. If this is not practical to see in deep learning, or if this functionality already exists please let me know!
For more context, see below where I have tried using EarlyStoppingHook.
UserWarning: Skip early stopping process since the evaluation results (dict_keys(['aAcc', 'mIoU', 'mAcc', 'mDice', 'mFscore', 'mPrecision', 'mRecall'])) do not include monitor (loss)
UserWarning: Skip early stopping process since the evaluation results (dict_keys(['aAcc', 'mIoU', 'mAcc', 'mDice', 'mFscore', 'mPrecision', 'mRecall'])) do not include monitor (val_loss)
Hi, as mentioned in #1486, mmengine doesn't support calculating the loss of a validation set yet, so early stopping can't perform logic based on the loss.
What is the feature?
Early stopping focused on the validation loss metric would be nice. From my minimal understanding, it would be hard to do with the current system of extending
BaseMetric
. At least with IoUMetric, there is no ability to early stop based on the validation loss. If this is not practical to see in deep learning, or if this functionality already exists please let me know!For more context, see below where I have tried using EarlyStoppingHook.
Any other context?
loss
early_stopping=dict(type='EarlyStoppingHook', monitor='loss', min_delta=0.01, patience=10)
monitor
(loss)val_loss
early_stopping=dict(type='EarlyStoppingHook', monitor='val_loss', min_delta=0.01, patience=10)
monitor
(val_loss)Related: