-
Hi, thanks for your great job. I found some errors when I wanted to continue my LoRA finetuning.
```
09/04/2024 15:54:23 - INFO - accelerate.accelerator - Loading states from output/tulu_v2_dolly_…
-
Hi Unsloth!
I came across this interesting model on reddit: https://www.reddit.com/r/LocalLLaMA/comments/1ez8rmu/llama31_just_got_ears_early_experiments/
It allows Text and Audio as input, and o…
-
Right now, when we finetune a LoRA on top of e.g. Llama 3.1 8B instruct, even if model_name is `meta-llama/Meta-Llama-3.1-8B-Instruct`, it gets resolved to `unsloth/meta-llama-3.1-8b-instruct-bnb-4bit…
-
### Your current environment
```text
The output of `python collect_env.py`
```
### How would you like to use vllm
Hi
I want to attach lora using docker command
docker run --runtime nv…
-
config file:
`# Model
pretrained_model_name_or_path = '/data/llm/cache/Qwen2-7B-Instruct/'
use_varlen_attn = True
# Data
data_files = ['/workspace/xtuner/sft_openai.json']
prompt_template = …
-
### Which Cloudflare product(s) does this pertain to?
Wrangler
### What version(s) of the tool(s) are you using?
Wrangler 3.72.2
### What version of Node are you using?
16.15.1
### W…
-
Trying to finetune unsloth/Mistral-Nemo-Instruct-2407-bnb-4bit. It gives me error with batch size 1 and max seq length of 2048.
I can see the example notebook on colab t4.
Unsloth: Fast Mistral pa…
-
Which one would you recommend to use to get the best possible performance?
-
I am trying to do a finetune using a custom dataset, in particular: https://huggingface.co/datasets/truthfulqa/truthful_qa
I haven't found any clear documentation, only partial docs explaining bits…
-
All of the models supported by torchtune currently have rather low context lengths (