-
### Question
ShareGPT is used for instruction fine-tuning, with the aim of inserting data from image independent pure text conversations into multiple rounds of image conversations, so that the model…
-
Thank you for your excellent work, but the open-source code indeed has many minor issues, which makes others hesitant to follow your work.
During the TrainStage1 phase, the issues are as follows:
1.…
-
[Recent discussion](https://discordapp.com/channels/1216089456134586388/1216091478581968958/1261063677365321771) with the community highlighted the interest in having a Navigation Engine able to choos…
-
When can we expect code update for GPU-based pre-training and fine-tuning instead of TPU? @ellisbrown @penghao-wu @tsb0601
-
Thanks for your great work!
I wanna know how you compute the raw token lens, just like the 729 in the image.
-
Thank you for sharing such an interesting idea!
> Since there is no longer a modality gap in the embeddings, we can transfer the single modality representation capabilities to multimodal embeddi…
-
您好,我想请问一下att_crop可以被应用到LLaVA上吗?
-
Hello, I've execute `main_qwen_npu` folloing the [guideline](https://github.com/UbiquitousLearning/mllm/tree/main/src/backends/qnn). In fact, there were minor bugs so I've manually fixed them. (e.g., …
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
github head
### Reproduction
```
--model_name_or_path /data/pretrained_models/MiniCPM-Llama3-V-2_5
-…
Vimos updated
4 weeks ago
-
I found with 2d5-7b the checkpoint saved from LoRA tuning finetune.py with one GPU is correct, while with multiple GPU the model saved is incorrect.
Does anyone met similar problem?
For example…