At the moment the Precision/Recall/Accuracy/F1-score calculation seems bizarrely wrong and I managed to reproduce that on a spreadsheet using what I understood to be the correct way when it comes to confusion matrix statistics in the multi-class setting.
Limdu's one seems to be in fact correct but it accounts for all labels predicted whereas I just look at the first one of in the arrays of predicted labels (assuming it's the most likely to be correct).
I don't know what exactly am I missing and if sticking to a proper multi-label (as opposed to just a multi-class) approach is better (even tho the learner will only care about one label per prediction).
At the moment the Precision/Recall/Accuracy/F1-score calculation seems bizarrely wrong and I managed to reproduce that on a spreadsheet using what I understood to be the correct way when it comes to confusion matrix statistics in the multi-class setting. Limdu's one seems to be in fact correct but it accounts for all labels predicted whereas I just look at the first one of in the arrays of predicted labels (assuming it's the most likely to be correct).
I don't know what exactly am I missing and if sticking to a proper multi-label (as opposed to just a multi-class) approach is better (even tho the learner will only care about one label per prediction).