H2O is an Open Source, Distributed, Fast & Scalable Machine Learning Platform: Deep Learning, Gradient Boosting (GBM) & XGBoost, Random Forest, Generalized Linear Modeling (GLM with Elastic Net), K-Means, PCA, Generalized Additive Models (GAM), RuleFit, Support Vector Machine (SVM), Stacked Ensembles, Automatic Machine Learning (AutoML), etc.
{noformat}searchParams.put("_booster", new XGBoostParameters.Booster[]{ //gblinear crashes currently
XGBoostParameters.Booster.gbtree, //default, let's use it more often
XGBoostParameters.Booster.gbtree,
XGBoostParameters.Booster.dart
});{noformat}
however, random search keeps track of indices combinations, not of parameters combinations... so that with the config above, it can produce twice the same model.
I guess the harm is not big today, as the hyperspace is large, but we need to reconsider this: I wish there was a way to give weights to some param, or maybe have a separate grid…
Need to figure out if it's worth having a full grid with similar search space for each booster though.
from {{XGBoostSteps}}:
{noformat}searchParams.put("_booster", new XGBoostParameters.Booster[]{ //gblinear crashes currently XGBoostParameters.Booster.gbtree, //default, let's use it more often XGBoostParameters.Booster.gbtree, XGBoostParameters.Booster.dart });{noformat}
however, random search keeps track of indices combinations, not of parameters combinations... so that with the config above, it can produce twice the same model. I guess the harm is not big today, as the hyperspace is large, but we need to reconsider this: I wish there was a way to give weights to some param, or maybe have a separate grid…
Need to figure out if it's worth having a full grid with similar search space for each booster though.