-
非常感谢作者~
我目前的情况是,当我使用8张GPU+deepspeed zero3+4bit qlora就会报错
和这个一样:https://github.com/microsoft/DeepSpeed/issues/3775
`RuntimeError: expected there to be only one unique element in `
在这个[讨论串](https://…
-
请问作者是否意愿实现qlora。
-
hello
when I train with multi gpu like this
```
WORLD_SIZE=8 CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 torchrun --nproc_per_node=8 qlora.py \
```
Then I get uneven VRAM utilization:
This me…
-
we are trying to finetune chatGLM6B using LoRA on arcA770 1card and 2cards , use the following command
1card:
```
python ./alpaca_lora_finetuning.py \
--base_model "/home/intel/models/chat…
-
Model: llama-2-7b-hf
Ubuntu :22.04
xpu-smi discovery:
uname -r
Steps followed:
1. Created conda env (followed instructions in Repo example)
2. Then initialized oneAPI
3…
-
Hi,
I have ported the Alpaca Qlora code given for GPU example, to CPU. I am using Sapphire Rapids for training.
These are my code changes:
model = AutoModelForCausalLM.from_pretrained( …
-
Hi,
It would be great to have an example of finetuning Phi without LoRA or QLoRA.
Thanks!
-
It looks like EleutherAI/gpt-j-6b is not supported:
Env:
Running from docker:
```
FROM pytorch/pytorch:2.0.1-cuda11.7-cudnn8-devel
RUN apt-get update && apt-get install git -y
RUN pip …
-
在用
torchrun --nproc_per_node=4 train.py --train_args_file train_args/sft/qlora/qwen2-7b-sft-qlora.json
训练qwen2+qlora+unsloth时(use_unsloth=true)出现错误:
ValueError: You can't train a model that has bee…
-
Hello,
I was trying to see if the Opacus library can be used with the Trainer module of huggingface. I see a code snippet in the readme that says to do the callback ```dp_transformers.PrivacyEngine…