chatchat-space / Langchain-Chatchat

Langchain-Chatchat(原Langchain-ChatGLM, Qwen 与 Llama 等)基于 Langchain 与 ChatGLM 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and Llama) RAG and Agent app with langchain
Apache License 2.0
29.96k stars 5.24k forks source link

[BUG] GLM4-Chat+Xinference:InternalServerError: Internal Server Error #4436

Closed WinstonCHEN1 closed 5 days ago

WinstonCHEN1 commented 1 week ago

问题描述 / Problem Description GLM4-Chat+Xinference,Follow了README的步骤,启动了模型之后,显存显示模型已经启动了,在xinference那边显示两个model也启动了。在chatchat的ui这边报错InternalServerError: Internal Server Error

环境信息 / Environment Information

附加信息 / Additional Information 添加与问题相关的任何其他信息 / Add any other information related to the issue. 以下是ui界面的报错信息 InternalServerError: Internal Server Error Traceback: File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/streamlit/runtime/scriptrunner/script_runner.py", line 600, in _run_script exec(code, module.dict) File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/chatchat/webui.py", line 69, in dialogue_page(api=api, is_lite=is_lite) File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/chatchat/webui_pages/dialogue/dialogue.py", line 361, in dialogue_page for d in client.chat.completions.create( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/_utils/_utils.py", line 277, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/resources/chat/completions.py", line 643, in create return self._post( ^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/_base_client.py", line 1261, in post return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/_base_client.py", line 942, in request return self._request( ^^^^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/_base_client.py", line 1026, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/_base_client.py", line 1074, in _retry_request return self._request( ^^^^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/_base_client.py", line 1026, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/_base_client.py", line 1074, in _retry_request return self._request( ^^^^^^^^^^^^^^ File "/home/chenguo/anaconda3/envs/chatchat/lib/python3.11/site-packages/openai/_base_client.py", line 1041, in _request raise self._make_status_error_from_response(err.response) from None image

EarZero commented 1 week ago

可以检查一下xinference服务是否可以正常推理,尝试使用命令行请求一下xinference上的模型。 类似: curl -X 'POST' \ 'http://127.0.0.1:9997/v1/chat/completions' \ -H 'accept: application/json' \ -H 'Content-Type: application/json' \ -d '{ "model": "my-llama-2", "messages": [ { "role": "system", "content": "You are a helpful assistant." }, { "role": "user", "content": "What is the largest animal?" } ] }'

tianxiajianghu commented 1 week ago

可以检查一下xinference服务是否可以正常推理,尝试使用命令行请求一下xinference上的模型。 类似: curl -X 'POST' 'http://127.0.0.1:9997/v1/chat/completions' -H 'accept: application/json' -H 'Content-Type: application/json' -d '{ "model": "my-llama-2", "messages": [ { "role": "system", "content": "You are a helpful assistant." }, { "role": "user", "content": "What is the largest animal?" } ] }'

单独运行上面的命令是可以正确回复的

zx1159652666 commented 1 week ago

有的模型不支持tool_choice=None的情况,如果为None,建议删掉,就可以正确回答了

jinceon commented 1 week ago

俺也一样

liunux4odoo commented 5 days ago

0.3.1 版已经发布,优化了配置方式,修改配置项无需重启服务器,可以更新尝试。

transformers 降级到 4.40 版本试试