Open ahmaurya opened 8 years ago
I'm guessing this might happen during the "Bayesian Optimization" stage but not during "Random Sampling" Stage. Because every time the algorithm think that point in your hyperparameter space is the point maximizing the utility function, the algorithm will keep choose that point for next round. My implementation will remove duplicated points for gaussian process to run in normal. A quick fix would be use one more round of random sampling when having this issue, but it would not be very elegant.
It would be great if you can share any solution about this. For now, maybe you can try matern 5/2 kernel or use a more aggressive parameter for your utility function. Good luck on Kaggle.
Perhaps you could delete from Mat_optim, params combination which are in DT_history, in line: argmax <- as.numeric(Mat_optim[which.min(Negetive_Utility), DT_bounds[, Parameter], with = FALSE]) before take the minimum. Other solution could be randomize the rounding to integer parameters when this happens. Instead of round to nearest integer random round with ceiling or floor.
I have the same problem, which package do you all switch to?
Hi Yachen,
I used the Bayesian Optimization package for optimizing the hyperparameters in a Kaggle contest. I noticed that the same hyperparameters can be repeatedly sampled, which can be a waste on really large datasets. Perhaps, it would be worthwhile to check if the hyperparameters have been already sampled and tried before? Thanks!