Open RonaldGalea opened 3 years ago
Have you tried removing all sampling hyperparameters from the search grid?
This is likely the same issue with global rng. I have a WIP branch removing it, will prioritize it.
@hcho3 Removing any of the 5 hyperparameters from the grid makes it work correctly. It is somehow these exact 5 that cause the issue.
Is this issue closed or is it still WIP?
Hello, I'm having an issue getting reproducible results when optimizing an
XGBClassifier
. I'm usingBayesSearchCV
to optimize some hyperparameters, defined in a search grid. When running locally, everything works as expected and I get the same results each time I run. However, running on a (local) dask cluster gives different results each time, please see the code snippet below.Code to reproduce issue:
Output:
It appears the differences come from the training of the
XGBClassifiers
, because the sets of evaluated hyperparameters are the same across runs. I also noticed there might be something amiss with this particular search grid, because if any of the entries are removed, the results are reproducible again.Library versions:
XGBoost version: 1.3.3 Scikit-optimize version: 0.9.dev0 Joblib version: 1.0.1 Dask version: 2021.04.0