-
Environment:
Docker version: 26.1.3
Nvidia-container-toolkit: 1.15.0-1
Inference: v0.11.2.post
cuda-driver: …
-
does spring ai support the vllm +qwen?
which starter can i use? and how to use spring ai connect the vllm ?
tks.
-
目前我们团队直接使用现在的代码对Qwen1.5的模型进行微调无法正常运行
-
- [ ] [Qwen-1.5-8x7B : r/LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/1atw4ud/qwen158x7b/)
# TITLE: Qwen-1.5-8x7B : r/LocalLLaMA
**DESCRIPTION:** "Qwen-1.5-8x7B
New Model
Someone creat…
-
### Your current environment
```text
The output of `python collect_env.py`
```
### 🐛 Describe the bug
python -m vllm.entrypoints.openai.api_server --model /root/autodl-tmp/models/Qwen1.5-14B-Ch…
-
如题,谢大佬回答
-
您好,您的工作确实很有创意。我剥离了原有的评估代码,使用Qwen1.5-14B-chat 进行测试。 当提供的文本长度达到30000字符长度时,InfLLM已经开始无法准确的回答问题。当文本继续增大时,输出错误加剧。能帮忙看看调整哪些参数可以优化结果
yaml 文件如下:
model:
type: inf-llm
path: /data/public/LLM/basemodel…
-
![微信图片_20240605135354](https://github.com/intel-analytics/ipex-llm/assets/166265863/4bcfc12a-ead8-468a-ab24-dfe60fb1d9d4)
The following error occurred after running for a period of time, please ref…
-
KwaiKEG/kagentlms_qwen_7b_mat
Qwen/Qwen-7B-Chat
两个都一样的提示。
尝试qwen1.5新版本,提示keyerror,qwen2
-
### Describe the bug
I found a bug is that the gpu memory shows the xinf and the llm model are running,but i can't see the model in web ui.Also,I can't use the llm model.
### To Reproduce
At begi…