Open kwen2501 opened 1 month ago
this highlights a current weakness - atm we are using the tokenizer downloaded by chat, but the safetensor weights by HF.
Thus, to get setup for a model we have to run:
1 - python3 torchchat.py download llama3
to get the tokenizer
2 - model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct")
to get the safetensor weights.
we can hook into the torchchat download.py to download the tokenizer, and then pull the safetensor weights via an HF call as well.
🐛 Describe the bug
Cc: @lessw2020
Versions
main branch