THUDM / GLM-4

GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型
Apache License 2.0
4.66k stars 371 forks source link

我将训练后的lora合并后,在执行新脚本时遇到了以下问题,请问这可能是什么原因导致的? #269

Closed Hyxxxxxx1 closed 2 months ago

Hyxxxxxx1 commented 2 months ago

Initializing an LLM engine (v0.4.3) with config: model='/jcdata/LLM/dev/chatglm4/merged_model', speculative_config=None, tokenizer=ChatGLM4Tokenizer(name_or_path='/jcdata/LLM/dev/chatglm4/glm-4-9b-chat', vocab_size=151329, model_max_length=128000, is_fast=False, padding_side='left', truncation_side='right', special_tokens={'eos_token': '<|endoftext|>', 'pad_token': '<|endoftext|>', 'additional_special_tokens': ['<|endoftext|>', '[MASK]', '[gMASK]', '[sMASK]', '', '', '<|system|>', '<|user|>', '<|assistant|>', '<|observation|>', '<|begin_of_image|>', '<|end_of_image|>', '<|begin_of_video|>', '<|end_of_video|>']}, clean_up_tokenization_spaces=False), added_tokens_decoder={ INFO 07-01 17:12:00 llm_engine.py:161] 151329: AddedToken("<|endoftext|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151330: AddedToken("[MASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151331: AddedToken("[gMASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151332: AddedToken("[sMASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151333: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151334: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151335: AddedToken("<|system|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151336: AddedToken("<|user|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151337: AddedToken("<|assistant|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151338: AddedToken("<|observation|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151339: AddedToken("<|begin_of_image|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151340: AddedToken("<|end_of_image|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151341: AddedToken("<|begin_of_video|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] 151342: AddedToken("<|end_of_video|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), INFO 07-01 17:12:00 llm_engine.py:161] }, skip_tokenizer_init=False, tokenizer_mode=auto, revision=None, rope_scaling=None, tokenizer_revision=None, trust_remote_code=True, dtype=torch.bfloat16, max_seq_len=8192, download_dir=None, load_format=LoadFormat.AUTO, tensor_parallel_size=1, disable_custom_all_reduce=False, quantization=None, enforce_eager=True, kv_cache_dtype=auto, quantization_param_path=None, device_config=cuda, decoding_config=DecodingConfig(guided_decoding_backend='outlines'), seed=0, served_model_name=/jcdata/LLM/dev/chatglm4/merged_model) Traceback (most recent call last): File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/transformers/utils/hub.py", line 398, in cached_file resolved_file = hf_hub_download( File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn validate_repo_id(arg_value) File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id raise HFValidationError( huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': 'ChatGLM4Tokenizer(name_or_path='/jcdata/LLM/dev/chatglm4/glm-4-9b-chat', vocab_size=151329, model_max_length=128000, is_fast=False, padding_side='left', truncation_side='right', special_tokens={'eos_token': '<|endoftext|>', 'pad_token': '<|endoftext|>', 'additional_special_tokens': ['<|endoftext|>', '[MASK]', '[gMASK]', '[sMASK]', '', '', '<|system|>', '<|user|>', '<|assistant|>', '<|observation|>', '<|begin_of_image|>', '<|end_of_image|>', '<|begin_of_video|>', '<|end_of_video|>']}, clean_up_tokenization_spaces=False), added_tokens_decoder={ 151329: AddedToken("<|endoftext|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151330: AddedToken("[MASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151331: AddedToken("[gMASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151332: AddedToken("[sMASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151333: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151334: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151335: AddedToken("<|system|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151336: AddedToken("<|user|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151337: AddedToken("<|assistant|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151338: AddedToken("<|observation|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151339: AddedToken("<|begin_of_image|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151340: AddedToken("<|end_of_image|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151341: AddedToken("<|begin_of_video|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151342: AddedToken("<|end_of_video|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), }'. Use repo_type argument if needed.

The above exception was the direct cause of the following exception:

Traceback (most recent call last): File "/jcdata/LLM/dev/chatglm4/basic_demo/openai_api_server_weitiao.py", line 670, in engine = AsyncLLMEngine.from_engine_args(engine_args) File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/vllm/engine/async_llm_engine.py", line 386, in from_engine_args engine = cls( File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/vllm/engine/async_llm_engine.py", line 340, in init self.engine = self._init_engine(*args, *kwargs) File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/vllm/engine/async_llm_engine.py", line 462, in _init_engine return engine_class(args, kwargs) File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/vllm/engine/llm_engine.py", line 212, in init self.tokenizer = self._init_tokenizer() File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/vllm/engine/llm_engine.py", line 408, in _init_tokenizer return get_tokenizer_group(self.parallel_config.tokenizer_pool_config, File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/vllm/transformers_utils/tokenizer_group/init.py", line 20, in get_tokenizer_group return TokenizerGroup(init_kwargs) File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/vllm/transformers_utils/tokenizer_group/tokenizer_group.py", line 23, in init self.tokenizer = get_tokenizer(self.tokenizer_id, tokenizer_config) File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/vllm/transformers_utils/tokenizer.py", line 92, in get_tokenizer tokenizer = AutoTokenizer.from_pretrained( File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 804, in from_pretrained tokenizer_config = get_tokenizer_config(pretrained_model_name_or_path, kwargs) File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 637, in get_tokenizer_config resolved_config_file = cached_file( File "/root/miniconda3/envs/chatglm4/lib/python3.10/site-packages/transformers/utils/hub.py", line 462, in cached_file raise EnvironmentError( OSError: Incorrect path_or_model_id: 'ChatGLM4Tokenizer(name_or_path='/jcdata/LLM/dev/chatglm4/glm-4-9b-chat', vocab_size=151329, model_max_length=128000, is_fast=False, padding_side='left', truncation_side='right', special_tokens={'eos_token': '<|endoftext|>', 'pad_token': '<|endoftext|>', 'additional_special_tokens': ['<|endoftext|>', '[MASK]', '[gMASK]', '[sMASK]', '', '', '<|system|>', '<|user|>', '<|assistant|>', '<|observation|>', '<|begin_of_image|>', '<|end_of_image|>', '<|begin_of_video|>', '<|end_of_video|>']}, clean_up_tokenization_spaces=False), added_tokens_decoder={ 151329: AddedToken("<|endoftext|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151330: AddedToken("[MASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151331: AddedToken("[gMASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151332: AddedToken("[sMASK]", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151333: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151334: AddedToken("", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151335: AddedToken("<|system|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151336: AddedToken("<|user|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151337: AddedToken("<|assistant|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151338: AddedToken("<|observation|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151339: AddedToken("<|begin_of_image|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151340: AddedToken("<|end_of_image|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151341: AddedToken("<|begin_of_video|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), 151342: AddedToken("<|end_of_video|>", rstrip=False, lstrip=False, single_word=False, normalized=False, special=True), }'. Please provide either the path to a local folder or the repo_id of a model on the Hub.

zRzRzRzRzRzRzR commented 2 months ago

2 请严格按照vLLM文档