Open konstantin-doncov opened 1 year ago
Sorry for the delay, I think the solution here is actually remove it all together. The logs say it only gets to try two models and both are worse than the dummy model, it seems like it needs to try more of them.
It could be that the logo
resampling strategy (which I guess is Leave One Group Out) could be creating many subsets of data which just means there's just too much data to fit if the number of groups is too much. Say for example you have 1_000_000
samples with 10 groups. My impression of logo
is that you would need to fit 10 models each on 900_000
equaling 9_000_000
data points total to get one model evaluation. This gets amplified more as the number of groups increases. Have you tried simple holdout just to test this hypothesis?
I want to use my own metric, but I get a lot of troubles during implementing this. Many of them the are related to each other. So, I hope I will solve all of them. E.g. if I use this code with 5 minutes max runtime(
time_left_for_this_task=5*60
):Then all fine and my metric function gets real predictions(not 0.5 0.5):
But if I use 4 minutes max runtime, then I get only dummy predictions(only 0.5 0.5).
You may say 'Well, then just use more time', but this is not a cure. Because when I use more complicated and time consuming (like 1-2 minute for one metric run) metrics, then it's not enough even one hour(and I don't know how much time it takes). So, how can I fix this?