This PR fixes the computation of positional encodings (PEs) for the full graph attention experiments, shown in the main paper (Table 1, Column 'Full Graph').
Due to the bug, for the full graph experiments, the PEs were computed on the fully connected (fully adjacent) graphs, and not the original sparse graphs.
With the correction, the PEs are calculated always on the original sparse graphs, which is the objective for PEs to capture original graph structure (hence positions as well) and inject them into the nodes.
--
Ps. Note that the full graph attention is not what the paper finds best for a graph transformer architecture, and this bug fix does not change the paper's main results, analysis and conclusion. The updated Table 1 will be on arxiv's next version of the paper.
This PR fixes the computation of positional encodings (PEs) for the full graph attention experiments, shown in the main paper (Table 1, Column 'Full Graph').
Due to the bug, for the full graph experiments, the PEs were computed on the fully connected (fully adjacent) graphs, and not the original sparse graphs. With the correction, the PEs are calculated always on the original sparse graphs, which is the objective for PEs to capture original graph structure (hence positions as well) and inject them into the nodes.
--
Ps. Note that the full graph attention is not what the paper finds best for a graph transformer architecture, and this bug fix does not change the paper's main results, analysis and conclusion. The updated Table 1 will be on arxiv's next version of the paper.
Thanks to @Saro00 for pointing this out.