graphdeeplearning / graphtransformer

Graph Transformer Architecture. Source code for "A Generalization of Transformer Networks to Graphs", DLG-AAAI'21.
https://arxiv.org/abs/2012.09699
MIT License
889 stars 137 forks source link

Fix in PEs compute for full graph experiments #2

Closed vijaydwivedi75 closed 3 years ago

vijaydwivedi75 commented 3 years ago

This PR fixes the computation of positional encodings (PEs) for the full graph attention experiments, shown in the main paper (Table 1, Column 'Full Graph').

Due to the bug, for the full graph experiments, the PEs were computed on the fully connected (fully adjacent) graphs, and not the original sparse graphs. With the correction, the PEs are calculated always on the original sparse graphs, which is the objective for PEs to capture original graph structure (hence positions as well) and inject them into the nodes.

--

Ps. Note that the full graph attention is not what the paper finds best for a graph transformer architecture, and this bug fix does not change the paper's main results, analysis and conclusion. The updated Table 1 will be on arxiv's next version of the paper.

Thanks to @Saro00 for pointing this out.