lm-sys / FastChat

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Apache License 2.0
36.63k stars 4.52k forks source link

fastchat-t5-3b-v1.0 模型无法本地运行, #1988

Open elmoss opened 1 year ago

elmoss commented 1 year ago

下载项目,下载依赖, 都是正常 首次运行项目: python3 -m fastchat.serve.cli --model-path lmsys/fastchat-t5-3b-v1.0 能正常运行, 文件缓存的目录: ~/.cache/huggingface/hub/models--lmsys--fastchat-t5-3b-v1.0

但是通过网页下载模型: https://huggingface.co/lmsys/fastchat-t5-3b-v1.0/tree/main 到vicuna_data目录,通过python3 -m fastchat.serve.cli --model-path vicuna_data/fastchat-t5-3b-v1.0 无法正常运行, 甚至将本地缓存的文件~/.cache/huggingface/hub/models--lmsys--fastchat-t5-3b-v1.0 一对一复制到vicuna_data/fastchat-t5-3b-v1.0目录也无法运行, 报错如下 Traceback (most recent call last): File "/root/anaconda3/lib/python3.9/runpy.py", line 197, in _run_module_as_main return _run_code(code, main_globals, None, File "/root/anaconda3/lib/python3.9/runpy.py", line 87, in _run_code exec(code, run_globals) File "/root/gpt/FastChat/fastchat/serve/cli.py", line 234, in main(args) File "/root/gpt/FastChat/fastchat/serve/cli.py", line 176, in main chat_loop( File "/root/gpt/FastChat/fastchat/serve/inference.py", line 261, in chat_loop model, tokenizer = load_model( File "/root/gpt/FastChat/fastchat/model/model_adapter.py", line 201, in load_model model, tokenizer = adapter.load_model(model_path, kwargs) File "/root/gpt/FastChat/fastchat/model/model_adapter.py", line 341, in load_model tokenizer = AutoTokenizer.from_pretrained( File "/root/anaconda3/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 702, in from_pretrained return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, *kwargs) File "/root/anaconda3/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1811, in from_pretrained return cls._from_pretrained( File "/root/anaconda3/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1965, in _from_pretrained tokenizer = cls(init_inputs, **init_kwargs) File "/root/anaconda3/lib/python3.9/site-packages/transformers/models/t5/tokenization_t5.py", line 154, in init self.sp_model.Load(vocab_file) File "/root/anaconda3/lib/python3.9/site-packages/sentencepiece/init.py", line 905, in Load return self.LoadFromFile(model_file) File "/root/anaconda3/lib/python3.9/site-packages/sentencepiece/init.py", line 310, in LoadFromFile return _sentencepiece.SentencePieceProcessor_LoadFromFile(self, arg) RuntimeError: Internal: src/sentencepiece_processor.cc(1101) [model_proto->ParseFromArray(serialized.data(), serialized.size())]

还有其他地方需要设置吗? 或者文件没下载完整?

PS: 按照官网示例: python3 -m fastchat.serve.cli --model-path lmsys/vicuna-7b-v1.3 下载vicuna-7b-v1.3模型到本地一样报错, 有谁遇到同样问题吗?

slievrly commented 1 year ago

放到 ~/.cache/huggingface/hub/models--lmsys--fastchat-t5-3b-v1.0/ 目录