-
### Your current environment
[Performance] 100% performance drop using multiple lora vs no lora(qwen-chat model)
gpu: 4 * T4
vllm version: v0.5.4
model:qwenhalf-14b-chat
### Model Input Dumps…
-
## 🐛 Bug
qwen-audio + vad 运行报错
### To Reproduce
1. Run cmd `python qwen_demo.py`
2. See error
```
2024-05-14 11:09:35,110 - modelscope - INFO - PyTorch version 2.3.0 Found.
2024-05-14 11:…
-
qwen的chat.so编不过,缺头文件和so,需要从sophon-demo的MiniCPM目录去取
-
### Your current environment
```text
PyTorch version: 2.2.0a0+git8964477
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.4 LTS (x86_64)
…
-
### Your current environment
```text
docker run --rm --runtime nvidia --gpus all --name vllm-qwen72b -v ~/.cache/huggingface:/root/.cache/huggingface \
-v /data1/Download/models/Qwen-7…
-
OSError: ./checkpoints/llava-qwen-4b-finetune-lora/ does not appear to have a file named config.json. Checkout 'https://huggingface.co/./checkpoints/llava-qwen-4b-finetune-lora//main' for available fi…
-
Hello,
I tried modifying the Qwen VL from this page to load your qwen_vl_guidance for inference but got the following error.
Traceback (most recent call last):
File "/workspace/dev/qwenvl/qwenv…
-
在运行rkllm时,user内容输入后,robot回答报错:
robot: :0: GGML_ASSERT(view_src == NULL || data_size == 0 || data_size + view_offs
-
How to configure the Hugging Face API? There is no Hugging Face option in the UI.
This configuration doesn't work.
openhands:
image: docker.all-hands.dev/all-hands-ai/openhands:0.13
…
-
Thanks!