Open Ronald-Kray opened 3 years ago
I usually use:
@WongKinYiu
I usually use:
- training data for training.
- validation data for choosing hyper-parameters.
- testing data for calculate mAP.
You mean that you spilt data Training data: Validation data: Testing data: 60:20:20 in the object detection problem? but, we can calculate mAP with only validation data (e.g.Training data: Validation date= 80:20). testing data for calculate mAP is necessary?
Because validation data is used for choosing the best model, that means the training may over-fit the validation data (this case is usually occur on the dataset which has only train/validation set or which provide ground truth for testing set). Only metrics on testing set without participate in developing step is objective results.
@WongKinYiu What about a custom dataset? Generally for papers on custom datasets, they only present mAP on Validation set in object detection problems. Is it wrong?
For custom dataset, it is hard to show your train/val/test set are separated well. In this case, you could show the mean and variance accuracy of cross validation results. Leave one out, jack knife, ... are often used methods.
@AlexeyAB @WongKinYiu I'm working on calculating mAP of the Object detection algorithms(Yolv5, Yolov4, EfficientDet).
In my opinion, there are 2 ways to calculate mAP(Assume that dataset is all labeled image).
1. Split dataset ratio as Train: Val=80:20, and just finish mAP calculation on Val dataset.
My question is that when calculating official mAP, Researchers use only a validation dataset? Or they calculate mAP of the test dataset from the model contributed by the validation dataset?