Open gptzerozero opened 1 year ago
@gptzerozero Thanks for asking. I was planning to clarify some of the differences in the readme.md but haven't gotten around to yet. This repository is inspired by alpaca-lora and mimics some of its features or limitations. One of the future goals is to implement linear modules if there is enough demand for it. However, this repository is already a good resource for anyone who wants to experiment with different LLMs and get the most out of the consumer hardware.
I notice that there are some differences compared to the
artido/qlora
repo. Why were the following code left out in this repo?https://github.com/artidoro/qlora/blob/3da535abdfaa29a2d0757eab0971664ed2cd97e8/qlora.py#L221-L232
and
https://github.com/artidoro/qlora/blob/3da535abdfaa29a2d0757eab0971664ed2cd97e8/qlora.py#L334-L344
Instead, you simply set the lora target modules to be
["q_proj", "v_proj"]
.Thank you for any clarifications