shaofanqi / CHATGLM4_Server

adapt Chatglm3 to load glm-4-9b-chat (Tool has some bugs)
7 stars 1 forks source link

由于用不了VLLM,想用这个脚本来跑GLM4但是报错 #1

Open Song345381185 opened 4 weeks ago

Song345381185 commented 4 weeks ago

服务可以起来但是调用就报错。

Song345381185 commented 4 weeks ago

服务可以起来但是调用就报错。

ERROR: Exception in ASGI application Traceback (most recent call last): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 401, in run_asgi result = await app( # type: ignore[func-returns-value] File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in call return await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in call await super().call(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/applications.py", line 123, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in call raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in call await self.app(scope, receive, _send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/cors.py", line 85, in call await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in call await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 756, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 776, in app await route.handle(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 77, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 72, in app response = await func(request) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 278, in app raw_response = await run_endpoint_function( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 191, in run_endpoint_function return await dependant.call(values) File "/workspace/GLM-4-zys/basic_demo/hf-api_server.py", line 315, in create_chat_completion response = generate_chatglm3(model, tokenizer, gen_params) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 263, in generate_chatglm3 for response in generate_stream_chatglm3(model, tokenizer, params): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 179, in generate_stream_chatglm3 for total_ids in model.stream_generate(inputs, eos_token_id=eos_token_id, **gen_kwargs): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/root/.cache/huggingface/modules/transformers_modules/modeling_chatglm.py", line 1206, in stream_generate logits_processor = self._get_logits_processor( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/transformers/generation/utils.py", line 871, in _get_logits_processor and generation_config._eos_token_tensor is not None AttributeError: 'GenerationConfig' object has no attribute '_eos_token_tensor'

shaofanqi commented 4 weeks ago

请问版本是啥呀

获取 Outlook for iOShttps://aka.ms/o0ukef


发件人: illusoryPython @.> 发送时间: Thursday, August 22, 2024 11:33:50 AM 收件人: shaofanqi/CHATGLM4_Server @.> 抄送: Subscribed @.***> 主题: Re: [shaofanqi/CHATGLM4_Server] 由于用不了VLLM,想用这个脚本来跑GLM4但是报错 (Issue #1)

服务可以起来但是调用就报错。

ERROR: Exception in ASGI application Traceback (most recent call last): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 401, in run_asgi result = await app( # type: ignore[func-returns-value] File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in call return await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in call await super().call(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/applications.py", line 123, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in call raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in call await self.app(scope, receive, _send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/cors.py", line 85, in call await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in call await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 756, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 776, in app await route.handle(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 77, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 72, in app response = await func(request) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 278, in app raw_response = await run_endpoint_function( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 191, in run_endpoint_function return await dependant.call(values) File "/workspace/GLM-4-zys/basic_demo/hf-api_server.py", line 315, in create_chat_completion response = generate_chatglm3(model, tokenizer, gen_params) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 263, in generate_chatglm3 for response in generate_stream_chatglm3(model, tokenizer, params): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 179, in generate_stream_chatglm3 for total_ids in model.stream_generate(inputs, eos_token_id=eos_token_id, **gen_kwargs): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/root/.cache/huggingface/modules/transformers_modules/modeling_chatglm.py", line 1206, in stream_generate logits_processor = self._get_logits_processor( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/transformers/generation/utils.py", line 871, in _get_logits_processor and generation_config._eos_token_tensor is not None AttributeError: 'GenerationConfig' object has no attribute '_eos_token_tensor'

― Reply to this email directly, view it on GitHubhttps://github.com/shaofanqi/CHATGLM4_Server/issues/1#issuecomment-2303607164, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABR5BUTUAIMHFZQGSSHXDZ3ZSVLZ5AVCNFSM6AAAAABM5H3VS2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMBTGYYDOMJWGQ. You are receiving this because you are subscribed to this thread.Message ID: @.***>

Song345381185 commented 4 weeks ago

请问版本是啥呀 获取 Outlook for iOShttps://aka.ms/o0ukef ____ 发件人: illusoryPython @.> 发送时间: Thursday, August 22, 2024 11:33:50 AM 收件人: shaofanqi/CHATGLM4_Server @.> 抄送: Subscribed @.*> 主题: Re: [shaofanqi/CHATGLM4_Server] 由于用不了VLLM,想用这个脚本来跑GLM4但是报错 (Issue #1) 服务可以起来但是调用就报错。 ERROR: Exception in ASGI application Traceback (most recent call last): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 401, in run_asgi result = await app( # type: ignore[func-returns-value] File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in call return await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in call await super().call(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/applications.py", line 123, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in call raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in call await self.app(scope, receive, _send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/cors.py", line 85, in call await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in call await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 756, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 776, in app await route.handle(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 77, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 72, in app response = await func(request) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 278, in app raw_response = await run_endpoint_function( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 191, in run_endpoint_function return await dependant.call(values) File "/workspace/GLM-4-zys/basic_demo/hf-api_server.py", line 315, in create_chat_completion response = generate_chatglm3(model, tokenizer, gen_params) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 263, in generate_chatglm3 for response in generate_stream_chatglm3(model, tokenizer, params): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 179, in generate_stream_chatglm3 for total_ids in model.stream_generate(inputs, eos_token_id=eos_token_id, *gen_kwargs): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/root/.cache/huggingface/modules/transformers_modules/modeling_chatglm.py", line 1206, in stream_generate logits_processor = self._get_logits_processor( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/transformers/generation/utils.py", line 871, in _get_logits_processor and generation_config._eos_token_tensor is not None AttributeError: 'GenerationConfig' object has no attribute '_eos_token_tensor' ― Reply to this email directly, view it on GitHub<#1 (comment)>, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABR5BUTUAIMHFZQGSSHXDZ3ZSVLZ5AVCNFSM6AAAAABM5H3VS2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMBTGYYDOMJWGQ. You are receiving this because you are subscribed to this thread.Message ID: @.>

请问版本是啥呀 获取 Outlook for iOShttps://aka.ms/o0ukef ____ 发件人: illusoryPython @.> 发送时间: Thursday, August 22, 2024 11:33:50 AM 收件人: shaofanqi/CHATGLM4_Server @.> 抄送: Subscribed @.*> 主题: Re: [shaofanqi/CHATGLM4_Server] 由于用不了VLLM,想用这个脚本来跑GLM4但是报错 (Issue #1) 服务可以起来但是调用就报错。 ERROR: Exception in ASGI application Traceback (most recent call last): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 401, in run_asgi result = await app( # type: ignore[func-returns-value] File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in call return await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in call await super().call(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/applications.py", line 123, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in call raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in call await self.app(scope, receive, _send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/cors.py", line 85, in call await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in call await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 756, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 776, in app await route.handle(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 77, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 72, in app response = await func(request) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 278, in app raw_response = await run_endpoint_function( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 191, in run_endpoint_function return await dependant.call(values) File "/workspace/GLM-4-zys/basic_demo/hf-api_server.py", line 315, in create_chat_completion response = generate_chatglm3(model, tokenizer, gen_params) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 263, in generate_chatglm3 for response in generate_stream_chatglm3(model, tokenizer, params): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 179, in generate_stream_chatglm3 for total_ids in model.stream_generate(inputs, eos_token_id=eos_token_id, *gen_kwargs): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/root/.cache/huggingface/modules/transformers_modules/modeling_chatglm.py", line 1206, in stream_generate logits_processor = self._get_logits_processor( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/transformers/generation/utils.py", line 871, in _get_logits_processor and generation_config._eos_token_tensor is not None AttributeError: 'GenerationConfig' object has no attribute '_eos_token_tensor' ― Reply to this email directly, view it on GitHub<#1 (comment)>, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABR5BUTUAIMHFZQGSSHXDZ3ZSVLZ5AVCNFSM6AAAAABM5H3VS2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMBTGYYDOMJWGQ. You are receiving this because you are subscribed to this thread.Message ID: @.>

您说的版本是啥?模型是GLM4-chat,python是3.10

shaofanqi commented 2 weeks ago

不好意思,HW机器封了。 应该是transformers的版本。 这边 Name: transformers Version: 4.41.2


From: illusoryPython @.> Sent: Wednesday, August 21, 2024 9:38 PM To: shaofanqi/CHATGLM4_Server @.> Cc: shaofanqi @.>; Comment @.> Subject: Re: [shaofanqi/CHATGLM4_Server] 由于用不了VLLM,想用这个脚本来跑GLM4但是报错 (Issue #1)

请问版本是啥呀 获取 Outlook for iOShttps://aka.ms/o0ukef … ____ 发件人: illusoryPython @.> 发送时间: Thursday, August 22, 2024 11:33:50 AM 收件人: shaofanqi/CHATGLM4_Server @.> 抄送: Subscribed @.> 主题: Re: [shaofanqi/CHATGLM4_Server] 由于用不了VLLM,想用这个脚本来跑GLM4但是报错 (Issue #1https://github.com/shaofanqi/CHATGLM4_Server/issues/1) 服务可以起来但是调用就报错。 ERROR: Exception in ASGI application Traceback (most recent call last): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 401, in run_asgi result = await app( # type: ignore[func-returns-value] File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in call return await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in call await super().call(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/applications.py", line 123, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in call raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in call await self.app(scope, receive, _send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/cors.py", line 85, in call await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in call await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 756, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 776, in app await route.handle(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 77, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 72, in app response = await func(request) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 278, in app raw_response = await run_endpoint_function( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 191, in run_endpoint_function return await dependant.call(values) File "/workspace/GLM-4-zys/basic_demo/hf-api_server.py", line 315, in create_chat_completion response = generate_chatglm3(model, tokenizer, gen_params) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 263, in generate_chatglm3 for response in generate_stream_chatglm3(model, tokenizer, params): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 179, in generate_stream_chatglm3 for total_ids in model.stream_generate(inputs, eos_token_id=eos_token_id, gen_kwargs): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/root/.cache/huggingface/modules/transformers_modules/modeling_chatglm.py", line 1206, in stream_generate logits_processor = self._get_logits_processor( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/transformers/generation/utils.py", line 871, in _get_logits_processor and generation_config._eos_token_tensor is not None AttributeError: 'GenerationConfig' object has no attribute '_eos_token_tensor' D Reply to this email directly, view it on GitHub<#1 (comment)https://github.com/shaofanqi/CHATGLM4_Server/issues/1#issuecomment-2303607164>, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABR5BUTUAIMHFZQGSSHXDZ3ZSVLZ5AVCNFSM6AAAAABM5H3VS2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMBTGYYDOMJWGQ. You are receiving this because you are subscribed to this thread.Message ID: @.>

请问版本是啥呀 获取 Outlook for iOShttps://aka.ms/o0ukef … ____ 发件人: illusoryPython @.> 发送时间: Thursday, August 22, 2024 11:33:50 AM 收件人: shaofanqi/CHATGLM4_Server @.> 抄送: Subscribed @.> 主题: Re: [shaofanqi/CHATGLM4_Server] 由于用不了VLLM,想用这个脚本来跑GLM4但是报错 (Issue #1https://github.com/shaofanqi/CHATGLM4_Server/issues/1) 服务可以起来但是调用就报错。 ERROR: Exception in ASGI application Traceback (most recent call last): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 401, in run_asgi result = await app( # type: ignore[func-returns-value] File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in call return await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in call await super().call(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/applications.py", line 123, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in call raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in call await self.app(scope, receive, _send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/cors.py", line 85, in call await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in call await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 756, in call await self.middleware_stack(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 776, in app await route.handle(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle await self.app(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 77, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/starlette/routing.py", line 72, in app response = await func(request) File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 278, in app raw_response = await run_endpoint_function( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/fastapi/routing.py", line 191, in run_endpoint_function return await dependant.call(values) File "/workspace/GLM-4-zys/basic_demo/hf-api_server.py", line 315, in create_chat_completion response = generate_chatglm3(model, tokenizer, gen_params) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 263, in generate_chatglm3 for response in generate_stream_chatglm3(model, tokenizer, params): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/workspace/GLM-4-zys/basic_demo/utils.py", line 179, in generate_stream_chatglm3 for total_ids in model.stream_generate(inputs, eos_token_id=eos_token_id, gen_kwargs): File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 36, in generator_context response = gen.send(None) File "/root/.cache/huggingface/modules/transformers_modules/modeling_chatglm.py", line 1206, in stream_generate logits_processor = self._get_logits_processor( File "/workspace/projects/GLM-4/glm4t4/lib/python3.10/site-packages/transformers/generation/utils.py", line 871, in _get_logits_processor and generation_config._eos_token_tensor is not None AttributeError: 'GenerationConfig' object has no attribute '_eos_token_tensor' D Reply to this email directly, view it on GitHub<#1 (comment)https://github.com/shaofanqi/CHATGLM4_Server/issues/1#issuecomment-2303607164>, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABR5BUTUAIMHFZQGSSHXDZ3ZSVLZ5AVCNFSM6AAAAABM5H3VS2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMBTGYYDOMJWGQ. You are receiving this because you are subscribed to this thread.Message ID: @.>

您说的版本是啥?模型是GLM4-chat,python是3.10

― Reply to this email directly, view it on GitHubhttps://github.com/shaofanqi/CHATGLM4_Server/issues/1#issuecomment-2303610948, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABR5BUTFDEGNPGPNEAY6VOTZSVML5AVCNFSM6AAAAABM5H3VS2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGMBTGYYTAOJUHA. You are receiving this because you commented.Message ID: @.***>