Open forceshorty opened 1 year ago
https://github.com/HuangLK/llama-deepspeed/blob/faedea514b11c18c695e1b2a6adb63b102ef001c/scripts/convert2ckpt.py#L65
Here is the hard code to add pad_token
Thank you for your answer. There is another question: why was not vocab_size increased by 1 in the convert2hf.py script, the original vocab_size is being used? https://github.com/HuangLK/llama-deepspeed/blob/faedea514b11c18c695e1b2a6adb63b102ef001c/scripts/convert2hf.py#L43
https://github.com/HuangLK/llama-deepspeed/blob/faedea514b11c18c695e1b2a6adb63b102ef001c/models/llama_pipeline_model.py#LL159C33-L159C33