-
### System Info
x86
Tensorrt_LLM 0.16.0
### Who can help?
_No response_
### Information
- [x] The official example scripts
- [ ] My own modified scripts
### Tasks
- [x] An officially supported …
-
-
Anyone successful finetune Qwen2-VL-7B-Instruct. I finetuned with 2 card A00 and got error CUDA OOM. Can you share which resources recommend to finetune it?
Thank you!
-
### System Info
qwen2-vl added new features of M-ROPE, please support it
### Who can help?
_No response_
### Information
- [ ] The official example scripts
- [ ] My own modified scripts
### Task…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
None
### Reproduction
lora qwen2-vl
### Expected behavior
Using Lora to train qwen2-vl, will the visu…
-
SOTA light weight vision model
[https://github.com/QwenLM/Qwen2-VL](https://github.com/QwenLM/Qwen2-VL)
llama.cpp issue [#9246](https://github.com/ggerganov/llama.cpp/issues/9246)
-
Can you please provide your inference code for Qwen2-VL-7B model. I am getting only 41.3% for the standard-4 choices case.
Below is my inference code.
-
对应的 gpu 信息如下:
![image](https://github.com/user-attachments/assets/1720ef39-51ef-4769-8408-e9c9482de9e8)
使用的是 docker,用的 paddle 镜像:ccr-2vdh3abv-pub.cnc.bj.baidubce.com/paddlepaddle/paddl:3.0.0b2-gpu-c…
luhui updated
6 hours ago
-
Hello
Will it be possible to include support for Qwen2-VL model? Thank you
-
### Motivation
1、The qwen2vl effect is the sota level in the open source model
2、lmdeploy is an excellent inference framework
3、So it's important to support turbomind
### Related resources
_No re…