-
Thank you for your great job. I met sorry wrong answer. Could you please tell me how to fix this?
![image](https://github.com/Pelochus/ezrknpu/assets/71435435/3e597dfe-02f6-44b4-9b37-5a9e04e31942)
…
udrs updated
4 months ago
-
### 是否已存在现有反馈与解答?
- [X] 我确认没有已有issue或discussion,且已阅读**常见问题**。
### 是否是一个代理配置相关的疑问?
- [X] 我确认这不是一个代理配置相关的疑问。
### 错误描述
不能推理
### 复现操作
1. 模型是qwen:14b-chat-v1.5-fp16
2. 运行ollama …
-
### Your current environment
```PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.4 LTS (x86_64)
GCC version:…
-
functions: Inputted functions for function calling. OpenAI format supported.
看到有这一条
请问他是指a list of tool 函数() 这种?但里面是不是还得改写?像 run 改成 call ? 还是可以整个 tool function直接用?
bzr1 updated
5 months ago
-
AssertionError
Traceback (most recent call last):
File "/Users/luxi.lm/opt/miniconda3/envs/Etuning/lib/python3.8/site-packages/gradio/routes.py", line 488, in run_predict
output = await app.g…
-
Here in git it says that it uses gpt 4o, but when testing the tool, it is at 3.5 turbo. Do I need to configure something to use 4o? Thank you very much!
-
[{'role': 'assistant', 'content': "I'm sorry, but as an AI language model, I don't have real-time access to current weather information. However, you can easily check the weather in San Francisco by s…
-
Hello,
I am trying to test Qwen/Qwen-7B-Chat with openai.api_server and this error shows up:
'''
vllm.engine.async_llm_engine.AsyncEngineDeadError: Task finished unexpectedly. This should never …
-
### Describe the bug
llama.cpp models always gives exactly the same output (compared in winmerge to be sure), like they ignore any sampling options and seed. Sometimes the first output after loadin…
-
VLLM 0.6.2 had just released few hours ago, it said no support multi image inference with Qwen2-VL.
I've try it, but it require the newest transformer and automatic install it.
When I start it u…