-
请问微调的qwen-7b是用仓库中的Evaluating-filtering-coling24-main/Qwen_finetune进行微调的吗?
我使用llm_batch_inference.py进行推理后的结果测试F1只有15.58,远低于论文中的42.94。
如果可以,请问是否能够给出lora模型checkpoint?
-
![1721889849832](https://github.com/user-attachments/assets/7d9b57ed-74cf-4edd-b6aa-61adf2f1bece)
mergekit-moe支持qwen系列的模型吗,我想把两个基于lora微调并合并后的qwen模型与没有微调的qwen合并为moe模型,但是我尝试了qwen、qwen1.5和qwen2三个版本,都得到图…
-
According https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard Qwen1.5 model is one of the best OpenSource (Free) models with large context and Russian language support. It would be nice to …
-
-
### Has this been supported or requested before?
- [X] I have checked [the GitHub README](https://github.com/QwenLM/Qwen2.5).
- [X] I have checked [the Qwen documentation](https://qwen.readthedocs.io…
-
### System Info
- GPU: NVIDIA H100 80G
- TensorRT-LLM branch main
- TensorRT-LLM commit: 8681b3a4c0ccc1028bb48d83aacbb690af8f55e7
### Who can help?
@byshiue @juney-nvidia @ncomly-nvidia
### Infor…
-
## My device information
```
NVIDIA Jetson AGX Orin Developer Kit(base) 64G
Package: nvidia-jetpack
Version: 6.1+b123
Priority: standard
Section: metapackages
Source: nvidia-jetpack (6.1)
Ma…
ljwps updated
1 month ago
-
Any plans to support quantitative reasoning for qwen models?
-
**Description**
I want to use VLMs with pytriton and vllm backend. Currently I am using sample script given at
https://github.com/triton-inference-server/pytriton/blob/main/examples/vllm/server.py
…
-
### Your current environment
```text
My environment:
vllm: 0.6.2
transformers: 4.45.2
auto_gptq: 0.7.1
```
### How would you like to use vllm
Hello, I have a question. I am deploying the Qwe…