ja-thomas / autoxgboost

autoxgboost - Automatic tuning and fitting of xgboost
Other
122 stars 19 forks source link

tune class weights param for unbalanced classif #35

Closed ja-thomas closed 6 years ago

ja-thomas commented 6 years ago

We're supporting weights, so class weights might be a nice way to improve the performance on unbalanced data.

But at some point it will get annoying with more and more wrapper, so we should def. consider mlrCPO

ja-thomas commented 6 years ago

We can just use the scale_pos_weight parameter

ja-thomas commented 6 years ago

done

p-schaefer commented 4 years ago

Unless I am mistaken, scale_pos_weight is only for binary classification, is there an alternative for multiclass classifications? Perhaps with xgboost.DMatrix(..., weight = *weight array for individual weights*), or passing .weights to trainLearner.classif.xgboost, or perhaps there is a better alternative?