Closed rohan-gt closed 3 years ago
Do you see the same behavior on RandomizedSearchCV?
No. Only on Bayesian and friends
@rohan-gt we can prioritize this in the upcoming week. thanks!
One possible solution could be if the user sets 10 trials, the first one could be using the algorithm's default parameters and the rest 9 could be sampled randomly and progress from there
Or the defaults could just be ran first, always - that way we guarantee the tuned model will not be worse than the one that went in.
What you want to do here is to implement something like Scikit-Optimize, where you can specify x0 and y0 upfront:
This probably needs to be handled outside the main optimization loop as it would be hard to just drop in the current model params into distributions (though it wouldn't be an issue with random/grid search).
While testing on multiple datasets I've observed that if
n_iter
is low, the tuned model has lower performance for both train and test sets compared to the model using default parameters. Is there a way to ensure that the default parameters of a model are tried out as the first trial and it starts optimizing from there? If there's no improvement, there's no need to run all the trials