dvas0004 / NerdNotes

A collection of notes: things I'd like to remember while reading technical articles, technical questions I couldn't answer, and so on.
11 stars 0 forks source link

Importance of probability calibration for classifiers #74

Open dvas0004 opened 4 years ago

dvas0004 commented 4 years ago

Classifiers in this note refer to:

PS, in anomaly detection if the output of the confidence is sufficiently low, the datapoint can be considered an anomaly.

Both types of classifiers however sometimes skew the results of their confidence measures (1)(2)

In order to resolve this, probability calibration is sometimes necessary. Basically the output of each classifier's confidence is passed through a regressor that has been trained on the predicted vs actual confidence. In other words:

1. The classifier is trained as usual 2. For each output class, create a regressor (isotonic or normal) 3. For every training sample, pass it through the model, and for each output class record the predicted probability and the actual probability (which is usually 0, or 1) 4. Train the regressors created in step 2 on the data collected in step 3

dvas0004 commented 4 years ago

http://fastml.com/classifier-calibration-with-platts-scaling-and-isotonic-regression/