Closed YenCheHsiao closed 1 week ago
Are you using zero3? If so, this is normal. Using zero2 will display the parameters.
Are you using zero3? If so, this is normal. Using zero2 will display the parameters.
Yes, I am using zero3, the same as in the provided file. Does using zero3 prevent training on the LLM, or is it just a display issue?
This should be a display issue, but we recommend using zero2.
Using zero2 does display the LLM parameters. Thanks.
In the
scripts/train/finetune.sh
file,--tune_type_llm
is set tofull
. I expect that all parameters in the language model would be trainable. However, the output file indicates that only parameters in the connector are trainable, while all LLM parameters remain non-trainable.Could you please clarify why the LLM parameters are not set as trainable when --tune_type_llm is configured to full? Based on the output, it appears only the connector parameters are trainable, while the LLM parameters remain at zero, which seems unexpected with the full setting.
Output: