Open patrick-miller opened 7 years ago
@patrick-miller great issue. It's something we should consider to avoid false research findings. You touch on several partial solutions, which I think all have merit.
One more possibility would be to choose a more conservative regularization strength than the highest performing alpha
in cross validation. There's an open issue on such approaches to prevent hyperparameter overfitting. But it would also help address multiple comparisons. In glmnet, I've been satisfied with lambda.1se
to set λ (the equivalent to alpha
in sklearn). This uses a "one-standard-error" rule to select a stronger regularization strength.
Let's keep this issue on our minds going forward.
I'm still working my way through the paper published by @gwaygenomics, @allaway and @cgreene, but it made me think of an issue that I believe we should try to deal with in our final product. In the paper they had a specific hypothesis that they tested; however, we are going to provide people with the ability to test out hypotheses on thousands of different mutations.
There are some problems with this ability, such as non-response bias. There are bound to be many uninteresting results (AUROC = 0.5) for different genes that people will tend to glance over. I can very easily imagine a scenario where someone iterates through many different genes until they reach one where a model does a good job at predicting a mutation.
We could approach this issue in a few different ways:
I wanted to open this issue up so we can discuss the importance of the problem and possible solutions.