Closed siboehm closed 2 years ago
Similar to your grover embedding, pretrained gnns can now be used to compute embeddings. Workflow is similar to yours, using parquet for dumping. See here: https://github.com/theislab/chemical_CPA/blob/0226a49485b5977e07fdd923e679d241a25569f4/notebooks/embedding_pretrained_gnn.ipynb
Since we now store those embeddings, what is left now is a simple class that uses a simple linear layer to map the latent space of the respective model.
For this, we would have to adjust:
self.dosers
are ohe with the respective dose, cf. https://github.com/theislab/chemical_CPA/blob/f9ba2e78446f712fed1daf1153dc1bcbd53ff395/compert/model.py#L412I agree with the three thoughts. Ref 3.): Yes we have to change this, I've already adapted the SigmoidDoser, the MLP still needs to be changed.
Add a class that takes in a list of canonical SMILES and returns a fixed-sized embedding of the molecules (
torch.Embedding
).Interface