Hi, thanks for the detailed tutorial of LoRA-from-scratch which helped me to understand the underlying priciples of the LoRA technique. Are there keras implementation of from-scratch LoRA for the RoBERTa model ? I am implementing the replace_multihead_attention_recursion of LoraWrapperRoberta.py in keras. But I don't know the right and complete practice in keras. I replaced the code with keras apis. It reported some errors and I could not find a way out.
Its pretty much optimized to work with pytorch directly. I'm not sure how keras wraps around that or how it uses the code snippets. Did you use the pytorch backend?
Hi, thanks for the detailed tutorial of LoRA-from-scratch which helped me to understand the underlying priciples of the LoRA technique. Are there keras implementation of
from-scratch LoRA for the RoBERTa model
? I am implementing the replace_multihead_attention_recursion ofLoraWrapperRoberta.py
in keras. But I don't know the right and complete practice in keras. I replaced the code with keras apis. It reported some errors and I could not find a way out.