Open clementruhm opened 14 hours ago
On duplicates: Can I ask what search space you have? Duplicates occurring might make sense if the search space is a small categorical space (with finitely many possibilities), but it would be shocking if this occurred with a continuous / DOUBLE search space - i.e. floating point parameter values being exactly the same.
On failed trials: Yes, you would mark the trial as infeasible with infeasibility_reason
. But from your code snippet, you seem to be using the designer
in a custom loop (rather than using our client API) - can I ask the reason for this?
yes, its multiple categorical or discrete parameters. The number of combinations is rather large (up to 500k). nevertheless, it seems the it converges pretty fast and tends to start giving duplicates. Is there way to dynamically increase exploration?
I started with using client API, then switched to designer for no particular reason. it seems to be a bit faster. I do need a custom loop though, because measurement will be async: I create a trial and in some point in the future - result arrives. For now I am testing it in sync setup, so it should not really affect for this question
Hi!
For the default designers (GPUCBPEBandit and GPBandit) what is the recommended way to deal with duplicates? If the same set of parameters is suggested again? Should I: a) store metrics from previous runs and complete trials using cached metric? b) sample another trial to save time?
Is there a way to hop out cycle of the duplicated set of parameters being suggested? For example dynamically increase exploration?
Another question, for the trials that fail, how do I report back that the set of parameters is invalid? i do:
Is it a valid approach? Is there a better one?
Kind regards