-
I've been finetuning unsloth/Phi-3-mini-4k-instruct-bnb-4bit with a T4, which doesn't support flash attention, so I don't have it installed.
During evaluation, I've been running into the following …
-
Hi,
I have successfully done the cycle of
```
llama download --source huggingface --model-id Llama3.1-8B-Instruct --hf-token
llama stack build
llama stack configure testerx
llama stack run …
-
1. **support plan**
when it will release the version for supporting llava-llama3-70b?
meainwhile, will it will consider of supporting unofficial version like, using llm of llama3-120b?
huggin…
-
### Describe the issue
Hello
Looking at the dataset list, which dataset does the prompts with an empty model belong to?
For example:
"id": "wgByO4Y_0",
"model": "",
Thanks
-
I am conducting the instruction tuning of llama3_llava using the script on my own dataset
`NPROC_PER_NODE=${GPU_NUM} xtuner train llava_llama3_8b_instruct_full_clip_vit_large_p14_336_lora_e1_gpu8_f…
-
When running the `finetune.sh` script with my own dataset, I encountered the following error during checkpoint / saving the model:
```
RuntimeError: The weights trying to be saved contained shared t…
-
### Question
Many thanks to the authors for this very good work.
I'm trying to run the instruct tuning part on one 3090 GPU, using the int-8 mode:
```
python llava/train/train.py \
--…
-
It is really impressive when doing inference with Mistral 7B. Thank you so much for open source it.
May I kindly ask what kind of format is the best way to finetune the model?
I read some blog p…
-
### What happened?
Hey all, I wanted to report a segmentation fault issue with llama-speculative. I have never once gotten this executable to work; I don't believe it is my command, as I have tried c…
-
thanks to the author. When porting to other LLM,such as baichuan 13B or qwen, there are a lot of warnings when doing finetuning, but the dataset LLaVA-Instruct-150K seem to right, What could be the pr…