但是通过网页下载模型: https://huggingface.co/lmsys/fastchat-t5-3b-v1.0/tree/main 到vicuna_data目录,通过python3 -m fastchat.serve.cli --model-path vicuna_data/fastchat-t5-3b-v1.0 无法正常运行,
甚至将本地缓存的文件~/.cache/huggingface/hub/models--lmsys--fastchat-t5-3b-v1.0 一对一复制到vicuna_data/fastchat-t5-3b-v1.0目录也无法运行, 报错如下
Traceback (most recent call last):
File "/root/anaconda3/lib/python3.9/runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/root/anaconda3/lib/python3.9/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/root/gpt/FastChat/fastchat/serve/cli.py", line 234, in
main(args)
File "/root/gpt/FastChat/fastchat/serve/cli.py", line 176, in main
chat_loop(
File "/root/gpt/FastChat/fastchat/serve/inference.py", line 261, in chat_loop
model, tokenizer = load_model(
File "/root/gpt/FastChat/fastchat/model/model_adapter.py", line 201, in load_model
model, tokenizer = adapter.load_model(model_path, kwargs)
File "/root/gpt/FastChat/fastchat/model/model_adapter.py", line 341, in load_model
tokenizer = AutoTokenizer.from_pretrained(
File "/root/anaconda3/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 702, in from_pretrained
return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, *kwargs)
File "/root/anaconda3/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1811, in from_pretrained
return cls._from_pretrained(
File "/root/anaconda3/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1965, in _from_pretrained
tokenizer = cls(init_inputs, **init_kwargs)
File "/root/anaconda3/lib/python3.9/site-packages/transformers/models/t5/tokenization_t5.py", line 154, in init
self.sp_model.Load(vocab_file)
File "/root/anaconda3/lib/python3.9/site-packages/sentencepiece/init.py", line 905, in Load
return self.LoadFromFile(model_file)
File "/root/anaconda3/lib/python3.9/site-packages/sentencepiece/init.py", line 310, in LoadFromFile
return _sentencepiece.SentencePieceProcessor_LoadFromFile(self, arg)
RuntimeError: Internal: src/sentencepiece_processor.cc(1101) [model_proto->ParseFromArray(serialized.data(), serialized.size())]
下载项目,下载依赖, 都是正常 首次运行项目: python3 -m fastchat.serve.cli --model-path lmsys/fastchat-t5-3b-v1.0 能正常运行, 文件缓存的目录: ~/.cache/huggingface/hub/models--lmsys--fastchat-t5-3b-v1.0
但是通过网页下载模型: https://huggingface.co/lmsys/fastchat-t5-3b-v1.0/tree/main 到vicuna_data目录,通过python3 -m fastchat.serve.cli --model-path vicuna_data/fastchat-t5-3b-v1.0 无法正常运行, 甚至将本地缓存的文件~/.cache/huggingface/hub/models--lmsys--fastchat-t5-3b-v1.0 一对一复制到vicuna_data/fastchat-t5-3b-v1.0目录也无法运行, 报错如下 Traceback (most recent call last): File "/root/anaconda3/lib/python3.9/runpy.py", line 197, in _run_module_as_main return _run_code(code, main_globals, None, File "/root/anaconda3/lib/python3.9/runpy.py", line 87, in _run_code exec(code, run_globals) File "/root/gpt/FastChat/fastchat/serve/cli.py", line 234, in
main(args)
File "/root/gpt/FastChat/fastchat/serve/cli.py", line 176, in main
chat_loop(
File "/root/gpt/FastChat/fastchat/serve/inference.py", line 261, in chat_loop
model, tokenizer = load_model(
File "/root/gpt/FastChat/fastchat/model/model_adapter.py", line 201, in load_model
model, tokenizer = adapter.load_model(model_path, kwargs)
File "/root/gpt/FastChat/fastchat/model/model_adapter.py", line 341, in load_model
tokenizer = AutoTokenizer.from_pretrained(
File "/root/anaconda3/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 702, in from_pretrained
return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, *kwargs)
File "/root/anaconda3/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1811, in from_pretrained
return cls._from_pretrained(
File "/root/anaconda3/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1965, in _from_pretrained
tokenizer = cls(init_inputs, **init_kwargs)
File "/root/anaconda3/lib/python3.9/site-packages/transformers/models/t5/tokenization_t5.py", line 154, in init
self.sp_model.Load(vocab_file)
File "/root/anaconda3/lib/python3.9/site-packages/sentencepiece/init.py", line 905, in Load
return self.LoadFromFile(model_file)
File "/root/anaconda3/lib/python3.9/site-packages/sentencepiece/init.py", line 310, in LoadFromFile
return _sentencepiece.SentencePieceProcessor_LoadFromFile(self, arg)
RuntimeError: Internal: src/sentencepiece_processor.cc(1101) [model_proto->ParseFromArray(serialized.data(), serialized.size())]
还有其他地方需要设置吗? 或者文件没下载完整?
PS: 按照官网示例: python3 -m fastchat.serve.cli --model-path lmsys/vicuna-7b-v1.3 下载vicuna-7b-v1.3模型到本地一样报错, 有谁遇到同样问题吗?