here is my error when running the inference.py. I checked that there is no mention of llava_llama whatsoever in this git or in transformer git. So I am confused with this error.
thank you !
python inference.py
python inference.py
[2024-06-27 16:22:53,054] [INFO] [real_accelerator.py:161:get_accelerator] Setting ds_accelerator to cuda (auto detect)
2024-06-27 16:22:53 PID=809676 __init__.py:49 setup_logging() INFO → 'standard' logger initialized.
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
Traceback (most recent call last):
File "/data/share2/zw/miniconda3/envs/cambrian/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 1118, in from_pretrained
config_class = CONFIG_MAPPING[config_dict["model_type"]]
File "/data/share2/zw/miniconda3/envs/cambrian/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 813, in __getitem__
raise KeyError(key)
KeyError: 'llava_llama'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/data/share2/zw/githubs/cambrian/inference.py", line 65, in <module>
tokenizer, model, image_processor, context_len = load_pretrained_model(model_path, None, model_name)
File "/data/share2/zw/githubs/cambrian/cambrian/model/builder.py", line 146, in load_pretrained_model
model = AutoModelForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs)
File "/data/share2/zw/miniconda3/envs/cambrian/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 526, in from_pretrained
config, kwargs = AutoConfig.from_pretrained(
File "/data/share2/zw/miniconda3/envs/cambrian/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 1120, in from_pretrained
raise ValueError(
ValueError: The checkpoint you are trying to load has model type `llava_llama` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
here is my error when running the inference.py. I checked that there is no mention of llava_llama whatsoever in this git or in transformer git. So I am confused with this error. thank you !
python inference.py
python inference.py