-
林哥,我想问一下M1 16G 的MacBook Air,想用您的这套代码Finetune一个模型,把batchsize和layer都调很低,大概得训练多久呢?还是说直接跑不动呢?实在没钱换M3了。那要是租云服务器的Nvida的卡,您这套基于mac的代码还能跑通吗?
-
### Your current environment
Collecting environment information...
PyTorch version: 1.12.1+cu113
Is debug build: False
CUDA used to build PyTorch: 11.3
ROCM used to build PyTorch: N/A
OS: Ubun…
-
### 🚀 The feature, motivation and pitch
I am currently using qwen1.5-14b, and I'm not sure if the templates for qwen2 are compatible with qwen1.5.
### Alternatives
_No response_
### Additional con…
-
According https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard Qwen1.5 model is one of the best OpenSource (Free) models with large context and Russian language support. It would be nice to …
-
在搭建Qwen1.5,不知道P40还是2080怎么选?
-
### System Info
- `transformers` version: 4.43.4
- Platform: Linux-4.9.151-015.ali3000.alios7.x86_64-x86_64-with-glibc2.17
- Python version: 3.8.18
- Huggingface_hub version: 0.24.6
- Safetensors…
-
I fine-tuned the model based on the [Qwen/Qwen1.5-0.5B-Chat](https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat) model and then fused the models. The final output when reasoning with the mlx-lm model is as…
-
hi unslothai, i got different inference result when using unsloth, i'v tested qwen1.5-chat and tinyllama-chat and got same issue, generate by unsloth always get a bad result compare with transformers …
-
https://github.com/eatsky/pot-app-translate-plugin-qwen
-
### Your current environment
Hi,
I did a full SFT on Qwen 0.5 B model using LLama-Factory, during which I specified the template parameter. I'm a little confused on whether I should to use a te…