Open shunxing12345 opened 5 months ago
It seems that the name of the embedding in your model is not 'embed_tokens'. You can modify it to the name of the embedding layer in your model.
Thanks for your replay! I got an other problem I am trying to train an LLM which structure differs from LLaMA and Mixtral, should I change the code of cnet.py? It seems based on LLaMA
This is not necessary; EAGLE's structure is independent of the target model. You can use the same cnet.py, or you can try other structures as well.
Thanks!
I have a finetuned a 12B model, but I got the OOM ERROR in model, head, optimizer, train_loader, test_loader, scheduler = accelerator.prepare( model, head, optimizer, train_loader, test_loader, scheduler
. I have 8 40G-A100.
this is my train_config
this is my config.json
I noticed that your "n_layers" is set to 38, which makes your draft model very large. In EAGLE, the draft model consists of only one layer.
Hi, I have successfully trained an Auto-regression Head, but I encountered the following error during inference. https://github.com/SafeAILab/EAGLE/blob/main/eagle/modeling_eagle.py#L957 and here is the size of Tensor
Hi i want to add a model that has a different architecture from the LLaMA model. BUT when I was trying
accelerate launch -m --mixed_precision=bf16 eagle.train.main --tmpdir [path of data]\ --cpdir [path of checkpoints] -- configpath [path of config file]
I got the following ERROR