-
-
According to https://python.useinstructor.com/concepts/maybe/.
There is an issue with tool calling in a case like this:
```json
{
"messages": [
{
"role": "system",
"conten…
-
### Describe the bug
I'm using langfuse to monitor one of dify's applications
Traces can be monitored, total_tokens is also there
The model I'm using is ollama's qwen:14b, configured it on the pa…
-
Since OpenAI-Compatible API Server now supports image input, can it support these 2 models?
https://github.com/THUDM/CogVLM
https://github.com/QwenLM/Qwen-VL
-
I am using the qwen 72B model, and the specified --conv-template does not take effect. If the stop parameter is not specified when calling, the conversation will never end.
启动命令
```
CUDA_VISIBLE_…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
my36z updated
2 weeks ago
-
Hi all,
I'm trying to run inference on Unsloth finetuned models. I'm using llama.cpp with 2x RTX 4090 GPUs to benchmark the performance of `Qwen2.5-Coder-14B-Instruct-128K-GGUF/Qwen2.5-Coder-14B-In…
-
### 🥰 需求描述
已在变量中配置好模型+qwen-vl-plus@Alibaba,+qwen-vl-max@Alibaba
能正常显示模型,但是path不对
普通qwen模型调用path是v1/services/aigc/text-generation/generation
qwen VL的path是v1/services/aigc/multimodal-generation/ge…
-
-
This error happens when tried to run this command for install autodistill qwen :
pip3 install autodistill-qwen-vl