-
**Describe the bug**
4xA100 gpu fine-tuning llama-3.1-8b-instruct (also tried llama2-13b-ms, same error), cli
```
CUDA_VISIBLE_DEVICES=0,1,2,3 \
NPROC_PER_NODE=4 \
LOCAL_WORLD_SIZE=4 \
swift…
-
In this command:
python finetune_mind2web.py --data_dir path/to/Mind2Web/data --base_model codellama/CodeLlama-7b-Instruct-hf --cache_dir --lora_dir --top_k_elements 20
I am not able to understand…
-
Please correct me if I'm wrong, but it looks like the current examples for lora training all build a loss function around completion, which lines up with the lora example of using only the 'text' fiel…
-
Hi, i tried finetuning both llama 3.1-8b-instruct and llama 3-8b-instruct following the notebook you provided [here](https://colab.research.google.com/drive/1XamvWYinY6FOSX9GLvnqSjjsNflxdhNc?usp=shari…
-
I am trying to finetune LLaMA 3.2 11B Vision Instruct on text *only*, but according to Unsloth this model is not supported yet. Is there a plan to support this. This is what I get from the command out…
-
**Describe the bug**
What the bug is, and how to reproduce, better with screenshots(描述bug以及复现过程,最好有截图)
The channel dimension is ambiguous. Got image shape (3, 672, 3). Assuming channels are the firs…
-
Hi, thanks for this amazing project. I was trying to finetune the lora model for Llama3.2 Vision which works fine and saved a adapter_0.pt; Then I wanted to use this adapter checkpoint for inference i…
-
Hi @hyeonahkimm, @Leaveson,
I would like to reproduce the results for mTSP mentioned in the Table 1 and Part B.2 Equity-Transformer of the paper.
As far as i understand, to achieve the results, …
-
Replicating my issue from Discord here.
I'm following the provided notebook on my dataset, but it keeps throwing the following error:
Unsloth: Untrained tokens of [[128042, 128036]] found, but e…
-
I’m trying to fine-tune Phi 3.5 Vision using transformers. However, I’m running into an issue trying to save the model during or after training. See below for a minimal reproducible example.
My examp…