-
I just use this command to start the server
`CUDA_VISIBLE_DEVICES=0 python -m sglang.launch_server --model-path LLMs/Qwen-14B-Chat --port 30000 --trust-remote-code --stream-interval 1 --enable-flash…
-
**功能描述 / Feature Description**
加问题的向量化和排名靠前的本地向量模型
**解决的问题 / Problem Solved**
目前我用qwen-api对知识库进行向量化后,后续的提问并不会向量化(没有访问qwen-api),不清楚是怎么做到精准的向量比对和score排序的
**实现建议 / Implementation Suggestions**
创…
-
Not able to run taskweaver with locally hosted Qwen1.5-72B-Chat.
Taskweaver worked fine with Qwen-72B-Chat.
Getting error:
Exception: OpenAI API request was invalid: Error code: 400 - {'object': …
-
当前选用的模型为glm4-chat,对话时报错,详细的错误为:
```
INFO: 127.0.0.1:62988 - "GET /tools HTTP/1.1" 200 OK
2024-07-16 16:39:32,138 httpx 21692 INFO HTTP Request: GET http://127.0.0.1:7861/tools "H…
-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [X] 2. The bug has not been fixed in the latest version.
- [X] 3. Please note that if the bug-related iss…
-
**问题描述 / Problem Description**
用简洁明了的语言描述这个问题 / Describe the problem in a clear and concise manner.
上传文件 没反应
**复现问题的步骤 / Steps to Reproduce**
1. 执行 '...' / Run '...'
2. 点击 '...' / Click '..…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
transformers>=4.37.2
datasets>=2.14.3
accelerate>=0.27.2
peft>=0.10.0
trl>=0.8.1
gradio>=4.0.0…
-
I have 8 RTX 4090 GPUs. Can they support a 70B-int4 parameter model?
-
### Reminder
- [X] I have read the README and searched the existing issues.
### Reproduction
python src/api_demo.py \
--model_name_or_path /model/Qwen1.5-7B-Chat \
--adapter_name_or_path …
-
**问题描述 / Problem Description**
chatchat配置oneapi平台之后,多功能对话报错;本地测试http://127.0.0.1:7861/chat/chat/completions接口,功能是正常的;oneapi的http://127.0.0.1:3000/v1/chat/completions接口也是正常的
2024-07-17 23:49:51.950…