-
### Question
After finetuning using the my custorm data, the finetuned llava model is overfitting. In my experiments, I following the your instrcuction( cited in https://github.com/haotian-liu/L…
-
As titled. Thank you !!!
-
### Question
Hi, thanks on your great work!
I use the following command to fine-tune the LLaVA-7b model.
`$PYTHON --nnodes=1 --nproc_per_node=8 --master_port=25001 \
llava/train/train_…
-
When I load the model as following, throw the error: Cannot merge LORA layers when the model is loaded in 8-bit mode
How can I load model with 4bit when inferencing?
`
model_path = 'decapoda-resea…
-
# URL
- https://arxiv.org/abs/2308.06259
# Affiliations
- Xian Li, N/A
- Ping Yu, N/A
- Chunting Zhou, N/A
- Timo Schick, N/A
- Luke Zettlemoyer, N/A
- Omer Levy, N/A
- Jason Weston, N/A…
-
Hi there! StarCoder from [BigCode](https://huggingface.co/bigcode) was trained for this kind of tasks, so having some documentation/support for it would be great.
Very nice project btw :fire:
-
### Feature request / 功能建议
In the finetune dir,the pipeline is train Q-A dataset,
however,we need to finetune MiniCPM by Instruct tuning.
Is there any solutions?
-
### Question
Thank you for your great work!
I am trying to fine-tune llava-v1.6-mistral-7b on the provided GQA dataset, using the script `finetune_task_lora.sh`. However, the loss dosen't decrea…
-
I've been testing running various finetuned versions of supported models on GKE. However, it gets stuck on ` Using the Hugging Face API to retrieve tokenizer config`
This are the full logs
```…
-
**LocalAI version: 1.22.0**
**Environment, CPU architecture, OS, and Version:**
Linux namehere 5.19.0-46-generic #47~22.04.1-Ubuntu SMP PREEMPT_DYNAMIC Wed Jun 21 15:35:31 UTC 2 x86_64 x86_6…