Closed bkowshik closed 1 year ago
You can use open-llama 3B by using https://github.com/Lightning-AI/lit-gpt. Particularly, you want to follow this tutorial: https://github.com/Lightning-AI/lit-gpt/blob/main/tutorials/download_openllama.md
Thank you @carmocca that makes sense. 👍
Some more context on this here: https://github.com/llm-efficiency-challenge/neurips_llm_efficiency_challenge/issues/6
This doesn't work because the 3B model is not sharded and there is a step in the convert hf code that requires the existence of pytorch_model.bin.index.json
to figure out how the sharding works. Any ideas how to bypass that step?
NOTE: Newbie to llama here ... trying to get this running following the documentation.
Ref: https://github.com/openlm-research/open_llama
Don't see the llama configs for the 3B model here: https://github.com/Lightning-AI/lit-llama/blob/main/lit_llama/model.py#L41-L46
Get a standard
KeyError
when I run the command below: