Closed hvgazula closed 2 years ago
Should we cache all models using git clone
or should I create a separate list with models that need to be cloned? Thoughts? @VeritasJoker @miahong
For more info please visit https://huggingface.co/docs/transformers/installation#offline-mode
I'm ok with both methods. does git clone every model cause a longer time? but it is a one-time thing right?
Yes, one time thing.
Closing for the time caching with both available methods. Will move to git lfs clone
in the future.
What is the issue? For example, if I cache a model using
model = AutoModelForCausalLM.from_pretrained("EleutherAI/gpt-neox-20b")
and then trying to load it usingmodel = AutoModelForCausalLM.from_pretrained("EleutherAI/gpt-neox-20b", local_files_only=True)
raises an OSError.Other Models (for our use case):
For all the above models, cloning (using
git lfs clone
) the model repo works fine. However, doing this will lead to inconsistent methods (in scripts/tfsemb_download.py) for caching models. i.e.AutoModel*.from_pretrained
andgit clone model_name
. For more info about these two modes, please refer to any model card in hugging face