-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
llama-factory中的视频数据处理逻辑,包括ProcessorMinix的使用,和qwen2官方给的推理代码中视频处理逻辑不一致。qwen2的处理逻辑参照:https://github.com/Qwen…
-
### System Info
- Ubuntu 20.04
- NVIDIA A100
### Who can help?
@Tracin @kaiyux
### Information
- [X] The official example scripts
- [ ] My own modified scripts
### Tasks
- [ ] A…
-
如题,我之前运行run_server还好好的,今天就报了这个错误
(qwen) D:\idata\Qwen-Agent>python run_server.py --llm qwen2:7b --model_server http://localhost:11434/v1/ --api_key EMPTY
2024-07-17 16:10:53,837 - run_server.py - 10…
-
Emu3 is a good work, but i have some question on it.
The vocabulary size of Qwen is 152064 , while the codebook size of vision tokenizer employed in Emu3 is 32768
The addation is 184832, the vocab…
-
### System Info
- transformers version: 4.45.2
- datasets version: 3.0.1
- Platform: Linux-5.15.0-1070-aws-x86_64-with-glibc2.35
- Python version: 3.10.12
- Huggingface_hub version: 0.26.1
- S…
-
Hi – thank you for open-sourcing this research project! Do you have plans to release code for fine-tuning these models?
-
### The model to consider.
Lora support for Qwen/Qwen2-57B-A14B (Qwen2MoeForCausalLM)
### The closest model vllm already supports.
Qwen/Qwen2-72B
### What's your difficulty of supporting the mod…
-
Great Work! I'm interested in unsloth and may I use it to finetune MLLM like Qwen-Vl?
-
### Before submitting your bug report
- [X] I believe this is a bug. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [ ] I'm not able to find an [open issue](ht…
-
File "/data/zengzijian/Projects/LLaMA-Factory-202407/src/llamafactory/model/patcher.py", line 30, in
from .model_utils.longlora import configure_longlora
File "/data/zengzijian/Projects/LLaMA-Facto…