Closed siboehm closed 2 years ago
The reason I'd run 1) on just a single embedding is for computational reasons, and because else things get very complex. We could also run the sweep on 2 embeddings, but all 7 just seems like it's overkill.
If it turns out that the transfer learning doesn't help for improving Trapnell scores, then there's no use in implementing #62 for example.
This is not super straight forward, I think. If you train a trapnell model independent from lincs, just on its own, you would choose HVGs (these are the meaningful genes wrt to the dataset). Hence, #62 somewhat allows us to regulate the degree at which we lean towards lincs genes or trapnell genes. After all, good/ reasonable r2 scores are required and should be aquired also in the transfer learning case but, ultimately, we are interested in the experiments around #67.
Having said that, I would also prioritise #67 at the moment. This will help us more on further experiment design.
predicting the effects of drugs
I am not wuite sure if I understand the difference to #67 here. Are you thinking about a scalar value that encode the 'effect' as in L2 norm between condition and control?
Sounds good wrt experiment design, lets put together a yaml file later.
In terms of the experiment hierarchy I was thinking this in terms of possible outcomes (best to worst):
We've decided to:
We'll use the outcomes of (2) for comparing the embeddings (including Vanilla!). Optionally we can schedule more runs using the optimal parameters for each embedding, but using different seeds to get an estimate of the variance.
Closed by #79
I'm not happy with the results of the large sweep that we ran on LINCS. Mainly:
I'd do it like this:
I think it's important to get this right before we design too many other experiments. If it turns out that the transfer learning doesn't help for improving Trapnell scores, then there's no use in implementing #62 for example. We can then still runs the other experiments like #67 and hope to see improvements there. If that doesn't work either, we can check whether the model is at least useful for predicting the effects of drugs that it hasn't seen. That's just for covering our bases in the worst case, I think with some tweaking the Transfer learning will work.