Closed phucdoitoan closed 2 months ago
Hi,
I am sorry for the late reply. As this is not the repository from the original KGT5 paper, results are expected to be slightly different. However, results should definitely be better than the 0.254 you report. Typically, with a larger effective batch size KGT5 performs considerably better. I would propose to use a batch size of at least 128 or 256 per GPU with more GPUs.
python main.py dataset.name=wikidata5m_v3 dataset.v1=True train.batch_size=128
In case you still have reproduction problems, you can use the pretrained checkpoint, which was trained with the original repo.
# Make sure you have git-lfs installed (https://git-lfs.com)
git lfs install
git clone https://huggingface.co/apoorvumang/t5-small-wd5mv3-adafactor_82ep
python eval_legacy_huggingface.py -m t5-small-wd5mv3-adafactor_82ep --v1 --is_legacy
In case you are comparing the performance to other models, I would propose to additionally use the descriptions, as they are also used with SimKGC.
python main.py dataset.name=wikidata5m_v3 dataset.v1=True train.batch_size=128 descriptions.use=True
Hi Adrian,
Thank you a lot for your detailed reply. It helps a lot.
Hi, I tried to reproduce the result of KGT5 with the command:
python main.py dataset.name=wikidata5m_v3 dataset.v1=True
run on 2 GPU, with original config.yaml file. But the result was MRR ~ 0.254, which is far lower than that in the KGT5 paper ( MRR ~0.300)Can you tell me how to reproduce the KGT5's resutls?
Thank you a lot!