Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Multiple comparisons problems #83

Open
patrick-miller opened this issue Feb 14, 2017 · 1 comment
Open

Multiple comparisons problems #83

patrick-miller opened this issue Feb 14, 2017 · 1 comment

Comments

@patrick-miller
Copy link
Member

I'm still working my way through the paper published by @gwaygenomics, @allaway and @cgreene, but it made me think of an issue that I believe we should try to deal with in our final product. In the paper they had a specific hypothesis that they tested; however, we are going to provide people with the ability to test out hypotheses on thousands of different mutations.

There are some problems with this ability, such as non-response bias. There are bound to be many uninteresting results (AUROC = 0.5) for different genes that people will tend to glance over. I can very easily imagine a scenario where someone iterates through many different genes until they reach one where a model does a good job at predicting a mutation.

We could approach this issue in a few different ways:

  1. hold out some data for validation -- only to be used for publication
  2. apply some sort of correction (e.g. Bonferroni)
  3. place strong emphasis on effect sizes
  4. list a clear disclaimer

I wanted to open this issue up so we can discuss the importance of the problem and possible solutions.

@dhimmel
Copy link
Member

dhimmel commented Feb 15, 2017

@patrick-miller great issue. It's something we should consider to avoid false research findings. You touch on several partial solutions, which I think all have merit.

One more possibility would be to choose a more conservative regularization strength than the highest performing alpha in cross validation. There's an open issue on such approaches to prevent hyperparameter overfitting. But it would also help address multiple comparisons. In glmnet, I've been satisfied with lambda.1se to set λ (the equivalent to alpha in sklearn). This uses a "one-standard-error" rule to select a stronger regularization strength.

Let's keep this issue on our minds going forward.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants