Closed YerongLi closed 1 year ago
--model_name_or_path ./checkpoints/vicuna-7b-v1.1 \
I got errorYou are using a model of type llama to instantiate a model of type llava. This is not supported for all configurations of models and can yield errors.
You are using a model of type llama to instantiate a model of type llava. This is not supported for all configurations of models and can yield errors.
With this --model_name_or_path LLaVA-7B-v0 \
I got this error.
Traceback (most recent call last):
File "/scratch/yerong/LLaVA/llava/train/train_mem.py", line 13, in <module>
train()
File "/scratch/yerong/LLaVA/llava/train/train.py", line 656, in train
model = LlavaLlamaForCausalLM.from_pretrained(
File "/scratch/yerong/.conda/envs/llava/lib/python3.10/site-packages/transformers/modeling_utils.py", line 2643, in from_pretrained
) = cls._load_pretrained_model(
File "/scratch/yerong/.conda/envs/llava/lib/python3.10/site-packages/transformers/modeling_utils.py", line 2952, in _load_pretrained_model
state_dict = load_state_dict(shard_file)
File "/scratch/yerong/.conda/envs/llava/lib/python3.10/site-packages/transformers/modeling_utils.py", line 431, in load_state_dict
raise OSError(
OSError: Unable to load weights from pytorch checkpoint file for 'LLaVA-7B-v0/pytorch_model-00001-of-00002.bin' at 'LLaVA-7B-v0/pytorch_model-00001-of-00002.bin'. If you tried to load a PyTorch model from a TF 2.0 checkpoint, please set from_tf=True.
[2023-07-19 04:54:54,848] [INFO] [launch.py:428:sigkill_handler] Killing subprocess 49958
use --model_name_or_path ./checkpoints/vicuna-7b-v1.1
You are using a model of type llama to instantiate a model of type llava. This is not supported for all configurations of models and can yield errors.
This is expected
When did you clone our code?
I cloned the code base after 5/1/23
Describe the issue
Issue: scripts/deepspeed/finetune_lora.sh I think in training workflow
--model_name_or_path
should not be vicuna-7b right, it reports some error as You are using a model of type llama to instantiate a model of type llava. This is not supported for all configurations of models and can yield errors.I tried both
--model_name_or_path LLaVA-7B-v0 \
or--model_name_or_path ./checkpoints/vicuna-7b-v1.1 \
, neither works though Command:Log:
Screenshots: You may attach screenshots if it better explains the issue.