Open iawen opened 1 year ago
If you load the ckpt and tokenizer WizardLM/WizardCoder-15B-V1.0
from Huggingface, I think you can delete this code.
If you load the ckpt and tokenizer
WizardLM/WizardCoder-15B-V1.0
from Huggingface, I think you can delete this code.
thx!
What hardware configuration is required to continue fine-tuning on WizardLM/WizardCoder-15B-V1.0? Also, fine-tuning uses the default configuration file: deepspeed_config.json, but throws an error: Not found scheduler!
I added according to the arguments of Starcoder:
{
"scheduler": {
"type": "WarmupLR",
"params": {
"warmup_min_lr": "auto",
"warmup_max_lr": "auto",
"warmup_num_steps": "auto"
}
}
}
Isn't it correct?
We met the same error before. We installed deepspeed==0.9.2
and transformers==4.29.2
to fix this error. We train our models with 8 V100-32GB GPUs.
We met the same error before. We installed
deepspeed==0.9.2
andtransformers==4.29.2
to fix this error. We train our models with 8 V100-32GB GPUs.
Whether all weights can be finetune with 8 V100-32G?
How long did it take you to train? Also based on WizardLM/WizardCoder-15B-V1.0?
Can the models fit on 4 x V100 16GB GPUs?
We met the same error before. We installed
deepspeed==0.9.2
andtransformers==4.29.2
to fix this error. We train our models with 8 V100-32GB GPUs.
How many epoch did you do and how long was the training time ?
Can someone put a detailed list of instructions to fine tuning the model? Or maybe link some guide.
Thank you
Can someone put a detailed list of instructions to fine tuning the model? Or maybe link some guide.
Thank you
Yes, please can someone give some guidance on how to fine-tune?
Hello, I'm going to fine-tune on WizardCoder-15B-V1.0 , do I need this code: