-
### Your current environment
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu Jammy Jellyfish (development branch…
-
From Twitter - adding new tokens to Qwen don't work?
```python
# Add special tokens to the tokenizer
num_added_tokens = tokenizer.add_special_tokens({"additional_special_tokens": special_tokens})
…
-
좋은 작업 감사합니다.
Qwen/Qwen2.5-72B-Instruct으로 실행했을때 아래와 같은 문제가 있습니다.
![image](https://github.com/user-attachments/assets/1ecf1efd-85d7-42ee-8935-d700a7d6acf5)
-
RTX 4090 24G,
Qwen-7B-Chat
loads OK:
```
model_config = ModelConfig(lora_infos={
"lora_1": conf['lora_1'],
"lora_2": conf['lora_2'],
})
model = ModelFactory.from_huggingface(conf['b…
-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
-
### Model Series
Qwen2.5
### What are the models used?
Qwen/Qwen2.5-1.5B-Instruct
### What is the scenario where the problem happened?
[inference with] with [vllm]
### Is this a known issue?
- …
-
Hello , in the https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int4/blob/c34a4a91629f09f73a285f32dbd26106b033c654/config.json#L29 has mentioned the groupsize is 128 for 4bit or 8bit. So could y…
-
这个proj非常好用,感谢开发组。
但是在应用过程中发现本proj主要支持LLM,能否提供一个支持Qwen-VL的例子呢(API调用),非常感谢!
-
Please add Qwen/Qwen2-7B model to the neuron cache
-
对qwen-2.5扩充词表后loss飙升到2000,可以帮忙看下问题吗