This training went well and I confirmed that a checkpoint output of 76G, which is the same as codellama 34b, was generated.
Afterwards, when trying to load the generated model through LlamaForCausalLM, the following error occurred.
ValueError: Trying to set a tensor of shape torch.size([0]) in "weight" (which has shape torch.Size([32000, 8192])), this look incorrect.
Thank you very much for your code.
I performed train.py with the codellama 34b base.
This training went well and I confirmed that a checkpoint output of 76G, which is the same as codellama 34b, was generated. Afterwards, when trying to load the generated model through LlamaForCausalLM, the following error occurred.
Is there anything I missed or need to fix?