-
I would like to ask how can I use QLoRa or Parameter-Efficient Fine-Tuning thin a model does not register at Hugging face instead is Based on [OFA](https://github.com/OFA-Sys/OFA)
i am trying to Qu…
-
GPU硬件 4张2080Ti,单卡显存12G,指定单卡运行。
执行chatglm-6b微调
`CUDA_VISIBLE_DEVICES=0 python train_qlora.py \
--train_args_json chatGLM_6B_QLoRA.json \
--model_name_or_path /data/chatglm-6b \
--train_data_path d…
-
In the paper, it was mentioned that 48G of graphics memory can train 65B of LLaMA
```
We present QLORA, an efficient finetuning approach that reduces memory usage enough to finetune a 65B parameter …
-
如题
-
### Your current environment
```text
The output of `python collect_env.py`
```
Collecting environment information...
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTo…
-
Hi, I noticed that you tried scaled rope. How could i use the scaled code for training a qlora with this?
ghost updated
8 months ago
-
While running:
**python lolrl_qlora_llama_hh.py --sampling_strategy good_priority**
logs with error msg like below:
[2024-03-19 18:59:01,658] [INFO] [real_accelerator.py:133:get_accelerator] …
-
Thanks for the code,
Is there any tweak we can do to fit it in Google Colab (~15 GB Memory)
or should I just settle with QLoRA?
-
Thanks you ,My english is senior,excuse me.
when i run the CUDA_VISIBLE_DEVICES=1 python3 train_qlora.py --train_args_json chatGLM_6B_QLoRA.json --model_name_or_path /T106/chatGLM-6B-QLoRA-main/chatG…
-
### Please check that this issue hasn't been reported before.
- [X] I searched previous [Bug Reports](https://github.com/axolotl-ai-cloud/axolotl/labels/bug) didn't find any similar reports.
### Exp…