Closed ernestmordret closed 4 months ago
I think this is a transformers
library issue, not related to Evo
you're absolutely right! I finally managed to fix it by running the following script AFTER running git clone git@hf.co:togethercomputer/evo-1-131k-base
from transformers import AutoConfig, AutoModelForCausalLM
model_name = 'togethercomputer/evo-1-8k-base'
offline = False
use_cache = True
local_files_only = False
model_config = AutoConfig.from_pretrained(model_name, trust_remote_code=True, offline=offline, use_cache=use_cache, local_files_only=local_files_only)
model_config.use_cache = True
model = AutoModelForCausalLM.from_pretrained(
model_name,
config=model_config,
trust_remote_code=True,
offline=offline,
use_cache=use_cache,
local_files_only=local_files_only
)
model.save_pretrained('evo-1-8k-base')
Now it seems to work, looking forward to experiment with it!
Great!
Hi! Thank you very much for the model, the pre-print looks fantastic!
I'd like to use your Huggingface model on our A100 GPUs, but unfortunately we have to work in a firewalled environment and therefore it complicates everything. I'm allowed to access the internet through the "submit" machines, that do not have GPUs, and then I have to switch to offline GPU machines to run the model.
For now, I have attempted to split the process in 2: First, use the submit machine to download the model locally
And then switch to my GPU machine and load the model with something like
but this time it fails miserably and spits this error:
Any idea why this might be the case?