-
# Description
I write a inference script like this:
```python
import torch
from PIL import Image
import sys
sys.path.append('./')
from llava.constants import IMAGE_TOKEN_INDEX, DEFAULT_IMAG…
-
I'm running the Llama-3-Instruct-8B-SPPO-Iter3 model locally and am very impressed by the improved quality from the original model. I can't help but wonder what the results would be if this finetunin…
-
hello, nice work. could share the sft-dataset in hf?
-
![微信截图_20240713161048](https://github.com/user-attachments/assets/179a13fc-1dce-45d5-b803-69151cab8e56)
转换后的文件和说明中的不同
-
### Bug Description
I have tried different open source llms,as well as gpt 3.5 turbo;Except for gpt for all other llms i had various kinds of issues like coroutine,key error,etc (For majority of my …
-
https://huggingface.co/Kooten/DaringMaid-20B-V1.1
^ That model has much better results than the original DaringMaid 20b from Kooten in terms of sticking with context.
I have wanted to use a Q8_0 …
-
Hello,
I have a dataset in .jsonl format of the following format:
`{"messages": [{"role": "system", "content": "some system msg"}, {"role": "user", "content": "some user input"}, {"role": "assist…
-
Hi!
Trying to finetune a Mistral 7B v0.2 on my GTX 1080 Max-q.
Getting this error (using LLaMA Factory):
==((====))== Unsloth: Fast Mistral patching release 2024.4
\\ /| GPU: NVID…
hvico updated
3 months ago
-
**Describe**
Thank you for your team's contribution! I would like to fine-tune E5-mistral-7b-instruct for tasks that interest me. Do you have plans to open-source training code? Alternatively, are th…
-
hey您好,看到你在huggingface上做的 starcoder3b-instruct模型,试了下符合预期,想了解下您这边finetune的过程,可否分享下