Closed PipiZong closed 5 months ago
Hi, there. Thanks for your support of our work. I will try my best to answer your questions:
Thanks for your reply. Sorry I am still confused about the second question, should we provide adapter_model.bin or pytorch_model.bin as the model path in inference.py?
And I have one more question, I downloaded your processed data and found the train data size for ml-1m and bookcrossing is around 7w and 1.5w. However, in your paper, as you mentioned in Table 2, the train size is 256/1024 on bookcrossing while 8192/65536 on ml-1m. Does this mean you sampled 256/1024 from 1.5w?
Thanks!
The 'model_path' is the path to your LLM weight, while the 'resume_from_checkpoint' is the path to the lora adapter weight. In finetune.py, the saved 'adapter_model.bin' will be renamed as 'pytorch_model.bin' for convenience, but they are the same thing. You can use them in the way you like if you are sure that it's the lora adapter weight. By the way, if you only want to test the zero-shot capability, then the lora weight is not needed.
Thank you for your reply. So we need to set use_lora to 1 for inference to reproduce your results? In the readme, use_lora seems to be 0 in your comment line for inference.
Hi, thanks for sharing your work. I have two questions about your code:
If you can provide the model you have in the drive folder and complete the command line with the provided model name, that would be easier to reproduce. Thanks