Open rickywu opened 4 months ago
I changed qanything_kernel/connector/llm/llm_for_openai_api.py this file return default token then can run with ollama
But new problem is not retrieve anything from knowledge base, fully answer by LLM self
I changed qanything_kernel/connector/llm/llm_for_openai_api.py this file return default token then can run with ollama
But new problem is not retrieve anything from knowledge base, fully answer by LLM self
me too!
是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
该问题是否在FAQ中有解答? | Is there an existing answer for this in FAQ?
当前行为 | Current Behavior
integrate with ollama and qwen:14b got error
期望行为 | Expected Behavior
Should work with ollama
运行环境 | Environment
QAnything日志 | QAnything logs
[2024-05-20 10:47:20 +0800] [9399] [ERROR] Exception occurred while handling uri: 'http://172.17.22.174:8777/api/local_doc_qa/local_doc_chat' Traceback (most recent call last): File "handle_request", line 132, in handle_request "_asgi_lifespan", File "/opt/miniconda3/envs/qanything-python/lib/python3.10/site-packages/sanic/response/types.py", line 547, in stream await self.streaming_fn(self) File "/opt/QAnything/qanything_kernel/qanything_server/handler.py", line 398, in generate_answer async for resp, next_history in local_doc_qa.get_knowledge_based_answer(custom_prompt=custom_prompt, File "/opt/QAnything/qanything_kernel/core/local_doc_qa.py", line 267, in get_knowledge_based_answer source_documents = self.reprocess_source_documents(query=query, File "/opt/QAnything/qanything_kernel/core/local_doc_qa.py", line 192, in reprocess_source_documents query_token_num = self.llm.num_tokens_from_messages([query]) File "/opt/QAnything/qanything_kernel/connector/llm/llm_for_openai_api.py", line 107, in num_tokens_from_messages raise NotImplementedError( NotImplementedError: num_tokens_from_messages() is not implemented for model qwen:14b. See https://github.com/openai/openai-python/blob/main/chatml.md for information on how messages are converted to tokens.
复现方法 | Steps To Reproduce
ollama run qwen:14b
script:
-b 'http://172.17.22.174:11434/v1' -k 'ollama' -n 'qwen:14b' -l '4096'
run and chat with qanything
备注 | Anything else?
No response