Open bilelomrani1 opened 1 year ago
Thanks for the suggestion! Fine-tuning (and LoRA support) is already on our roadmap - We'll definitely be looking into this.
Thank you, that's great! Out of curiosity, is your roadmap public and visible somewhere?
It's not publicly available at the moment.
Low-Rank Adaptation (LoRA) has become the de-facto parameter-efficient finetuning technique to adapt a base language model to a specific task.
curated-transformers
already supports dynamic quantization usingbitsandbytes
, hence adding some utilities to inject trainable adapters opens the door to usingcurated-transformers
as a replacement to the HuggingFacetransformers
+peft
stack. This could also enable a very nice finetuning integration into spaCy in the future.For reference, I find this implementation in
lit-gpt
really readable.Do you find this idea interesting?
If so, as for the user-facing API, drawing inspiration from HuggingFace
peft
it could look something like