yuyangw / MolCLR

Implementation of MolCLR: "Molecular Contrastive Learning of Representations via Graph Neural Networks" in PyG.
MIT License
233 stars 57 forks source link

Reproducing benchmark results on the SIDER dataset #15

Closed icycookies closed 1 year ago

icycookies commented 1 year ago

Thanks for your inspiring work. However, we've got problems when fine-tuning MolCLR on the SIDER dataset. It's reported in the paper that MolCLR achieves 68.0 in terms of ROC_AUC under Scaffold split on SIDER, but we fail to reproduce the results. Here's what we've tried:

  1. In our implementation, the average ROC_AUC of MolCLR over 27 tasks using default settings is approximately 62.3, which is far from the reported results. I observe that in the original paper of MoleculeNet, models are evaluated using Random split instead of Scaffold split on SIDER. It seems that the comparison between RF (68.4) and MolCLR (68.0) in the benchmark table is unfair.
  2. We conduct experiments with Random split on SIDER. Here's the results.
屏幕快照 2022-09-21 下午2 41 54

We've also tried several hyper-parameter combinations by changing dropout, hidden size and activation functions of MLP, which yields similar results. We hope the authors could kindly offer the experiment settings and hyper-parameters on SIDER that fully reproduces the promising results of MolCLR.

yuyangw commented 1 year ago

Thanks for your interest in our work. A suggested setting for SIDER:

Also, tuning the hyperparameters for each task individually can improve performance. Hope this helps.

Best, Yuyang

icycookies commented 1 year ago

Thanks for your response! We've successfully reproduced the reported results on SIDER.