Closed Newiz430 closed 1 year ago
Sorry for the late reply; I was on vacation.
Sorry for the late reply; I was on vacation.
- Thank you for bringing that to our attention. We did not include the experiments on ogbl-collab initially. We will address the careless errors in our arXiv version.
- I recommend adding batch normalization to the encoder and using learnable embeddings as input node features.
- This learning process consists of two stages. First, we pre-train GraphMAE to obtain node representations. Then, we train the edge decoder using the learned representations for link prediction tasks.
Not the answer I was looking for about the 3rd question (I was asking the pre-training configs and workflow of GraphMAE without node labels, like if the edges were split during node attribute prediction pre-training, how did you validate the pre-trained model without using the original logistic regression classifier etc). Still thanks a lot for replying tho
Hi Dr. Li,
Loved your work that promotes the self-supervised masked structural modeling! I am currently reproducing your results reported in Table 3 and have several questions.
train_linkpred_ogb.py
? Section 6.1.2 in your paper stated that "For all datasets except arXiv, two GCN layers are applied for the encoder" but the default encoder for ogbl-collab is "sage" (train_linkpred_ogb.py
, line 99). Also, Section 6.1.2 in your paper stated that MaskGAE uses BCE as link prediction loss but the implementation uses some other loss called "auc" (train_linkpred_ogb.py
, line 212).GNNEncoder
withtorch_geometric.nn.models.VGAE
ortorch_geometric.nn.models.ARGVA
(with an extra 2-layer GCN for variance learning) would not really work because the encoder kept outputing large features & logits that lead to very large loss values and training failure. Considering your results in Table 3, I'm very grateful for your advice about the training settings of those variational autoencoders.I much appreciate your timely reply so that I can cite your paper! Thank you so much!