Hello author,
Thank you for your open source TinyLLama1B, but I tried to train tiny_LLaMA_1b based on your pretrain codebase. After loading the huggingface format model, when I imported the lit_gpt format, I found that the stata_dict keywords of the model did not match. One was in GPT format and the other was LlamaModel, as shown below:
Is there any good solution?
Looking forward to your reply.
Hello author, Thank you for your open source TinyLLama1B, but I tried to train tiny_LLaMA_1b based on your pretrain codebase. After loading the huggingface format model, when I imported the lit_gpt format, I found that the stata_dict keywords of the model did not match. One was in GPT format and the other was LlamaModel, as shown below:![image](https://github.com/jzhang38/TinyLlama/assets/35062202/af1f1eed-cdcd-49af-88a8-3f080bdf68de)
Is there any good solution?
Looking forward to your reply.