-
### Your current environment
```text
PyTorch version: 2.4.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.4 LTS (x86_64)
GCC ve…
-
配置文件内容如下:
`
# 如果你没有 OPENAI API Key,可以使用 g4f 代替,或者使用国内的 Moonshot API
llm_provider="openai" # "openai" or "moonshot" or "oneapi" or "g4f" or "azure"
########## OpenAI API Key
# …
-
### Summary
i get error after run command 👍
` wasmedge --dir .:. --nn-preload default:GGML:CPU:/root/.cache/nexa/hub/official/Qwen2.5-Coder-1.5B-Instruct/q4_0.gguf llama-api-server.wasm --pro…
-
Hi thanks for providing such wonderful evaluation toolkit.
I was wondering why evaluation on `mmlu_generative` returns 0 accuracy whenever what models I try (pythia, qwen).
I understand it as …
-
### 🥰 需求描述
由于o1-mini以及o1-preview模型的API接口不支持流式传输及temperature必须为1,无法调用此API模型,希望能够适配完成
### 🧐 解决方案
希望适配o1-mini以及o1-preview模型
### 📝 补充信息
![image](https://github.com/user-attachments/assets/96c55831-11…
-
我发现qwen的原生agent的react是放在了user上,而swift官方提供的ms_agent的react放在了system中,这两者有何区别,或者哪种更好?
Qwen:
```bash
system
You are a helpful assistant.
user
Answer the following questions as best you can. You …
-
非常赞的工作!有几个使用上的问题请教一下:
在一台物理机上(8张4090卡),我启动了一个容器,容器内可以使用这8张卡,如果使用vllm部署Qwen2-7B-Instruct实例,一个卡一个实例,我可以通过指定不同的端口来启动模型服务,比如 8001~8008,那么对应的我有8个模型服务;如果切换到 llumnix,作为我的调度器,这8个模型服务应该如何启动呢?
根据这里的描述,https…
-
This is a ticket to track a wishlist of items you wish LiteLLM had.
# **COMMENT BELOW 👇**
### With your request 🔥 - if we have any questions, we'll follow up in comments / via DMs
Respond …
-
### Self Checks
- [X] This is only for bug report, if you would like to ask a question, please head to [Discussions](https://github.com/langgenius/dify/discussions/categories/general).
- [X] I hav…
-
### Your current environment
```text
Collecting environment information...
WARNING 07-23 19:11:42 _custom_ops.py:14] Failed to import from vllm._C with ModuleNotFoundError("No module named 'vllm.…