hiyouga / LLaMA-Factory

Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
https://arxiv.org/abs/2403.13372
Apache License 2.0
34.26k stars 4.22k forks source link

How to continue training LoRA made without llama factory? #5848

Open Sehyo opened 2 weeks ago

Sehyo commented 2 weeks ago

Reminder

System Info

I have a lora adapter I have made before with unsloth. I want to do KTO training on it with Llama factory, how can I do this? Setting Custom model and lora adapter path does not seem to work. I would prefer not having to merge the model, as I want to run just the base model in vllm and then load a single lora. How can I accomplish this? Thanks!

Reproduction

(Help)

Expected behavior

No response

Others

No response

hiyouga commented 2 weeks ago

have you tried --adapter_name_or_path?

Sehyo commented 2 weeks ago

Hi there! How would I configure that with the webui? Thanks!