vksastry / gnn_anl_gc_collab

0 stars 2 forks source link

Improve network architecture flexibility #4

Open WardLT opened 2 years ago

WardLT commented 2 years ago

The current architecture is tuned for molecular energy by predicting the energy as a sum over all atoms. Our models from the older GNN implementation that worked for redox property best predict redox potential based on a whole-graph fingerprint

See our paper on GNNs for solvation energy. The text around Figure 1 explains how moving the readout function can have a large effect.

WardLT commented 2 years ago

Once done, evaluate performance on the redox dataset (#2)

WardLT commented 2 years ago

TL;DR: This issue is a low priority because the current network architecture works almost as well as my original.

Turns out my network from the original MPNN is only slightly better. I get an MAE of 0.133 V on the redoxmer dataset #5 compared to the 0.140 V I get using this MPNN.

That architecture using a reduce_sum to combine atomic fingerprints into a molecular fingerprint before passing them through the output MLP. The implementation in this repository does a reduce_sum afterwards, which does not seem to be needed in light of other architectural changes (e.g., this repo using "global" state and my original MPNN does not).