Closed lightest1 closed 3 days ago
Thank you for your question! Yes, in this case, LoRA is applied only to the linear layers in the FFN (gate_proj
, down_proj
, up_proj
). However, you can modify the code to include q_proj
, k_proj
, or v_proj
(which are related to the self-attention layer) as lora_targets
based on your specific needs.
Hello! Thank you for your work! From the code, it seems that lora only acts on gate_proj,down_proj and up_proj. I want to know whether the result in the paper is the result of these lora_targets, or whether the code should be modified into q_proj, k_proj or v_proj?