cognoma / machine-learning

Machine learning for Project Cognoma
Other
32 stars 47 forks source link

Multiple comparisons problems #83

Open patrick-miller opened 7 years ago

patrick-miller commented 7 years ago

I'm still working my way through the paper published by @gwaygenomics, @allaway and @cgreene, but it made me think of an issue that I believe we should try to deal with in our final product. In the paper they had a specific hypothesis that they tested; however, we are going to provide people with the ability to test out hypotheses on thousands of different mutations.

There are some problems with this ability, such as non-response bias. There are bound to be many uninteresting results (AUROC = 0.5) for different genes that people will tend to glance over. I can very easily imagine a scenario where someone iterates through many different genes until they reach one where a model does a good job at predicting a mutation.

We could approach this issue in a few different ways:

  1. hold out some data for validation -- only to be used for publication
  2. apply some sort of correction (e.g. Bonferroni)
  3. place strong emphasis on effect sizes
  4. list a clear disclaimer

I wanted to open this issue up so we can discuss the importance of the problem and possible solutions.

dhimmel commented 7 years ago

@patrick-miller great issue. It's something we should consider to avoid false research findings. You touch on several partial solutions, which I think all have merit.

One more possibility would be to choose a more conservative regularization strength than the highest performing alpha in cross validation. There's an open issue on such approaches to prevent hyperparameter overfitting. But it would also help address multiple comparisons. In glmnet, I've been satisfied with lambda.1se to set λ (the equivalent to alpha in sklearn). This uses a "one-standard-error" rule to select a stronger regularization strength.

Let's keep this issue on our minds going forward.