-
A: libc++abi: terminating due to uncaught exception of type std::out_of_range: unordered_map::at: key not found
[1] 58628 abort ./cli_demo ../qwen1.5-0.5b-chat
-
Hi,
What is the number of training tokens for the base models Qwen 1.5 0.5B, 1.8B, 4B, 7B, 14B, 72B,110B、Qwen2 7B 72B models?
Thanks for the great work!
-
### Describe the bug
I use docker-compose to deploy Xinference. Most of the time it works fine, but at some random moment, a KeyError is triggered, causing the entire service to fail. Here are my ste…
-
### System Info
transformers version : 4.38.1
platform: ubuntu 22.04
python version : 3.10.14
optimum version : 1.19.2
### Who can help?
@ArthurZucker and @younesbelkada
### Information
- [X] …
-
openai的模型可使用tiktoken这个包
请问qwen模型该如何统计流式输出的token数量呢?
`
llm2 = ChatOpenAI(
model="/vllm/qwen1.5-chat-moe",
openai_api_key = 'xxx',
openai_api_base…
-
### Your current environment
环境:
torch 2.3.0
vllm 0.5.0.post1
transformers 4.41.2
主要报错情况:
moe小一点的模型 '/data/models/qwen/qwen1.5-2.7Bmoe' 不会出问题
对于大一点的就报错如最下面。
代码:
from vllm.engine.arg_ut…
-
Hi folks, please report issues regarding the ragas custom critic model in this thread.
Docs : https://docs.ragas.io/en/latest/howtos/customisations/ragas_custom_model.html
Model: https://huggingfa…
-
- 系统: debian 12
- ollama: 0.1.40
- qwen1.5 正常, qwen2 不正常:
-
错误截图:
![图片](https://github.com/xorbitsai/inference/assets/29749635/27488b41-296f-4fe7-8c64-4aa0cf78cd03)
模型:Qwen1.5-14B-Chat-GPTQ-int4
加载引擎:vllm
错误信息:
torch.cuda.OutOfMemoryError: [address=0.0.0.…
-
请问下,能直接跑 Qwen1.5-72B-Chat-GPTQ-Int4 模型吗?