-
### Please check that this issue hasn't been reported before.
- [X] I searched previous [Bug Reports](https://github.com/axolotl-ai-cloud/axolotl/labels/bug) didn't find any similar reports.
### Exp…
-
如题,想知道8bit和4bit qlora效果有没有差别?
-
I used: train_qlora.py to fine-tuned the model for llama 2-7b, and then used get_predict_qlora.sh ( the checkpoint is 10000) to get the results, but many of the outputs are empty, as shown below:
…
-
我严格按照README安装了相关的包
pip install -r requirements.txt
pip install git+https://github.com/unslothai/unsloth.git
pip install bitsandbytes==0.43.1
pip install peft==0.10.0
pip install torch==2.2.2
pip…
-
### System Info
Latest TRL from source, can't run TRL env rn as cluster is shut down but I'm installing everything from source.
If required will restart cluster and run.
### Information
- [ ] Th…
-
Hi, thanks for the interesting project!
I create Gemma 7B based model [webbigdata/C3TR-Adapter](https://huggingface.co/webbigdata/C3TR-Adapter).
This model is Huggingface transformer format and …
-
Traceback (most recent call last):
File "/root/xx/DB-GPT-Hub/predict_qlora.py", line 233, in
dataset_name, result = predict()
File "/root/xx/DB-GPT-Hub/predict_qlora.py", line 109, in pred…
-
## Typology of Efficient Training
- Data & Model Parallel
- Data Parallel
- Tensor Parallel
- Pipeline Paralle
- Zero Redundancy Optimizer(ZeRO) (DeepSpeed, often work with CPU offloadi…
-
When I run the qlora example with the OneAPI 2024 installed, it prompts an error where libsycl.so.7 is not found.
```
warnings.warn(
0%| | 0/20…
-
we are trying to finetune chatGLM6B using LoRA on arcA770 1card and 2cards , use the following command
1card:
```
python ./alpaca_lora_finetuning.py \
--base_model "/home/intel/models/chat…