-
We try to implement 4bit-qlora, thanks to the optimized kernel implementation of back-propagation, the fine-tuning speed is similar to 8-bit lora at present. Welcome to use and issue: https://github.c…
-
-
qlora微调模型输出正确结果后仍然会输出一些不相干的内容,例如:
![image](https://github.com/yangjianxin1/Firefly/assets/59114904/e3b50b77-165b-4757-b4eb-0a6349ec1f12)
我使用的断句功能,但是他在断完句子后仍然会输出一些无关紧要的内容,我一开始以为是训练集的大小太小,于是我将训练集的大小从2…
-
When I use the command below I got an error:
```shell
python3 qlora.py –learning_rate 0.0001 --model_name_or_path
```
╭─────────────────────────────── Traceback (most recent call last) ─…
-
Hi,
I am trying to use the Qlora code as provided in the repo on a Sapphire Rapids, Flex GPU machine.
I was able to run the [qlora_finetuning.py](https://github.com/intel-analytics/BigDL/blob/m…
-
In the given examples axoltol [exmaples/medusa](https://github.com/ctlllll/axolotl/tree/main/examples/medusa),
I follow the `vicuna_7b_qlora_stage1.yml` and `vicuna_7b_qlora_stage2.yml` to write my …
-
> Today we’re releasing the next step: QDoRA. This is just as memory efficient and scalable as FSDP/QLoRA, and critically is also as accurate for continued pre-training as full weight training. We thi…
-
Right now, [`requirements.txt`](https://github.com/artidoro/qlora/blob/main/requirements.txt) has `accelerate @ git+https://github.com/huggingface/accelerate.git`, but as of now this breaks QLoRA func…
-
Qlora LLaMa 13B
```
File "/home/hysz/anaconda3/envs/qlora/lib/python3.10/site-packages/torch/optim/lr_scheduler.py", line 69, in wrapper
return wrapped(*args, **kwargs)
File "/home/hysz/…
-
Hi Team,
I have successfully finetuned a QLoRA adapter on a custom dataset. When I try to load it in full precision, it gets loaded and works well
But this takes too much time and GPU memory to …