Closed rytheranderson closed 10 months ago
Hi,
Thanks for your question! Please use "post_smiles" and "post_tanimoto" for evaluation.
"post_smiles" is derived by the model generated SMILES and postprocessing based on the predicted atoms and bonds.
"graph_smiles" is constructed entirely by the predicted atoms and bonds to form the graph.
In our experiments, we observe that "post_smiles" generally works better. Therefore we take it as our model's prediction.
Aha!
"post_smiles" is derived by the model generated SMILES and postprocessing based on the predicted atoms and bonds.
This is the piece I was missing, thanks for the quick and concise explanation.
Closing, as my question is resolved.
When retraining, what are the correct metrics to compare with the paper? After retraining the full model (1m680k) it looks like the best model "post_smiles" metric matches Table 2 closely (with some variations as I used a smaller batch size) and, similarly, "post_tanimoto" metric matches Table 2 in the SI. I want to make sure I have the correct perspective on the retraining process.
If you have the time, a brief description of each metric in:
would be greatly appreciated. I am familiar with tanimoto similarity, SMILES, etc. But, I am trying to gain a better understanding of how they are implemented here, and how the different score variations ("" vs. "post" vs. "graph") can be interpreted to better understand the model.
Thanks in advance, -Ryther