-
in app.py file, init meth setting model_name = "cambrian_qwen" and model_path = "./checkpoints/longvu_qwen"。 but README tell us to download models for LongVU_Qwen2_7B_img 、LongVU_Qwen2_7B or LongVU…
-
### Proposal to improve performance
vllm serve /workspace/model/llm/Qwen/Qwen2_5-3B-Instruct\
--host 0.0.0.0 \
--port 2017 \
--tensor-parallel-size 1 \
--gpu-memory-utilization …
-
Could you please support Qwen VL model
-
### System Info
- nvcr.io/nvidia/tritonserver:24.08-trtllm-python-py3
### Who can help?
@kaiyux How to use a simple method to prove that enable_kv_cache_reuse is working correctly?
### Information…
-
https://huggingface.co/collections/Qwen/qwen25-math-66eaa240a1b7d5ee65f1da3e
-
I use the same script as in the Readme:
```
accelerate launch --num_processes 8 --main_process_port 12345 -m lmms_eval \
--model longva \
--model_args pretrained=lmms-lab/LongVA-7B,conv_…
-
### Problem
Hello,
I'm very pleased to see exllama getting vision capabilities for the first time with Pixtral!
You hinted at supporting new models in the release notes. What models are you h…
-
### System Info
python==3.10.15
cuda==11.8-8.8.1
torch==2.4.0
The latest version of code
GPU A100_40G * 8
### Who can help?
@ziyuwan @Gebro13 @mengfn @gzqaq @YanSong97 @i
### Information
- …
-
作者你好,Codev-Bench是很赞的工作。
论文中Codev-Agent有4个步骤,其中会调用 Qwen 分析repo源码、readme、并解析代码AST等,最后编写单元测试。
请问相关步骤中对Qwen使用到的prompt会开源吗?感谢
-
Environment
• Docker Image: nvcr.io/nvidia/tritonserver:24.10-trtllm-python-py3
• TensorRT-LLM Version: 0.14.0
• Run Command:
python3 ../run.py \
--input_text "你好,请问你叫什么?" \
--max_output_len=…