Closed trollkotze closed 4 months ago
Okay, sorry for being lazy and stupid. Apparently one can just use local relative paths by prefixing with a './' and that's the just a methods from the transformers library. https://github.com/huggingface/transformers/issues/14336
I noticed that
AutoTokenizer.from_pretrained(model_name)
will always redownload the model from HF, which seems quite stupid and useless.I've no idea where the model goes when it's downloaded, but is there any way to just keep it? I don't want to redownload the same thing over and over again.
And if I want to load another model, if I replace
mistralai/Mistral-7B-Instruct-v0.1
with e.g.152334H/miqu-1-70b-sf
I get an error about a missing tokenizer, which isn't included in that repository. I think I can just use the llama2-70b tokenizer here, because that's what this model is based on. If I could just put it all in place locally instead of relying on a fresh download from an HF repo that needs to contain everything I could probably piece it together.