Closed pricexu closed 1 month ago
In addition, I saw this question shown solved here #11 but after I changed the code to the format as #11, the results are still the same (~0.16 on acc and ~0.36 on pubmed). I noticed that the accs are only problematic on the Cora and Pubmed so maybe the node orders in "cora.csv" and "pubmed.csv" are not the original orders from the OGB package?
Hi Zhe,
The mismatch might be due to the data loader. For Cora and PubMed, you should use the load_data function we provide: https://github.com/XiaoxinHe/TAPE/blob/26f1e43b6aa9de39a8f68ab79d1f2b607d8baf01/core/data_utils/load.py#L26, as the data ordering is not the same as in the original order from the OGB package.
Please try replacing the data loader, and feel free to let me know if that doesn't work. Thanks!
Thank you, Xiaoxin! It helps. :-)
Hi Xiaoxin,
Thanks for this interesting work. I was checking the acc of the LLM's prediction using your provided data in "gpt_preds" However, the acc I tested is different from what you replied. My following piece of code is used, where I copied your load_gpt_preds function:
and what I see is
I think the acc on ogbn-arxiv matches what you reported, but the accs on Cora and Pubmed are much lower than the numbers you said 0.6769 and 0.9342.
I am wondering if the file "cora.csv" and "pubmed.csv" is wrong , could you help have a very quick check? I would greatly appreciate that!
Zhe