Open guozhiyao opened 1 year ago
I change the 13b llm and pretrain the project layer by loading your stage2 checkpoint. The initial loss is around 5 and drop slowly. It is around 2.5 after 1.5m samples. Is it normal? I have try to fine tune your model with the same config and model as your stage2 and the loss is around 1.4. I do not know if my pretrain loss can match your training log or not.
Can you do fine-tune using 7B?
Can you do fine-tune using 7B?
@yPanStupidog I try to change the llm. But I do not know if my pretrain loss is normal. Because I found the fine tune loss of your 13b is around 1.4. I have not try the 7b yet. Could you provide the training log of 7b or 13b please?
Could you please help me check whether finetune could be applied on 7B?I always met errors!
13B stage2 finetune in my case, the initial loss is 3.8, epoch 4, 200 iters, the best loss is 1.9
Can you do fine-tune using 7B?
@yPanStupidog I try to change the llm. But I do not know if my pretrain loss is normal. Because I found the fine tune loss of your 13b is around 1.4. I have not try the 7b yet. Could you provide the training log of 7b or 13b please?
@guozhiyao a quick question, what's your config for pre-training? I tried to pre-train this model using the default training config (stage 1) with a 7b LLM, it gave me nan
loss at the beginning, did you meet the similar error? Thanks
Can you do fine-tune using 7B?
@yPanStupidog I try to change the llm. But I do not know if my pretrain loss is normal. Because I found the fine tune loss of your 13b is around 1.4. I have not try the 7b yet. Could you provide the training log of 7b or 13b please?
@guozhiyao a quick question, what's your config for pre-training? I tried to pre-train this model using the default training config (stage 1) with a 7b LLM, it gave me
nan
loss at the beginning, did you meet the similar error? Thanks
same error
Can you do fine-tune using 7B?
@yPanStupidog I try to change the llm. But I do not know if my pretrain loss is normal. Because I found the fine tune loss of your 13b is around 1.4. I have not try the 7b yet. Could you provide the training log of 7b or 13b please?
@guozhiyao a quick question, what's your config for pre-training? I tried to pre-train this model using the default training config (stage 1) with a 7b LLM, it gave me
nan
loss at the beginning, did you meet the similar error? Thankssame error
I have encountered the same problem. Have you resolved it
same issue, could you provide some insights?
I use vicuna7b and EVA01-CLIP-g-14 , the initial loss is 6.8479. After 1.0M sampls, the loss is 2.6, is it normal?
Could you provide the training log of stage1 and stage2?