-
Hi guys,
Now I can fine-tune 'visionbranch_stage2_finetune.yaml' on **four** A100 80GB GPUs using gradient accumulation.
I'd like to know at what point the Loss is considered to have converged? …
-
code
```python
'''
conda activate beyond_scale_2_unsloth
'''
import torch
from datasets import load_dataset
from trl import SFTConfig, SFTTrainer
from unsloth import FastLanguageModel
from tr…
-
-
Hi,
I am trying to reproduce the results of llama-adapter v2. I am finetuning the model with "alpaca_gpt4_data" and "llava_instruct_150k" datasets and using the settings from https://github.com/Ope…
-
![b642c6cb-0ea5-4f69-8fc9-a82dfaed1616](https://github.com/user-attachments/assets/b0c57787-e777-4e58-94ff-d834c2ef8273)
-
### 🐛 Describe the bug
I am currently running the Colossalai/examples/language/openmoe project with the following experimental setup:
datasets: load_dataset("yizhongw/self_instruct/data/finetuning…
-
### What happened?
When converting models using [convert_hf_to_gguf.py](https://github.com/ggerganov/llama.cpp/blob/master/convert_hf_to_gguf.py) to GGUF format, a `TypeError` occurs if the `licens…
-
### System Info
Name: peft
Version: 0.13.2
### Who can help?
when i try to load the adapter for inference its showing the following error.
`TypeError: LoraConfig.__init__() got an unexpected ke…
-
I've fine-tuned the **llama 3 8 billion** model. I followed the notebook and only changed the dataset. The dataset is similar to the **alpaca dataset** but for the **Bangla language**. I've trained th…
-
### What happened?
u0_a227@localhost ~> ./llama.cpp/build/bin/llama-cli -m llama.cpp/models/Qwen2.5-0.5B-Instruct-Q4_K_M.gguf -p "You are a helpful assistant" -cnv -ngl 99 -t 8 -b 64 -tb 8 --ctx-size…