-
**Describe the bug**
I was training to run sft based on Mixtral-8x7B-instruct model with tensor parallel size=4 (sequence parallel=True) and LoRA (target modules =[all]).
It reports that the output …
-
Hi, Thank you for your work. I am trying to run train_text_to_image_lora.py on my own dataset, but I encountered the following error:
`ValueError: The deprecation tuple ('LoRAAttnProcessor', '0.30.…
-
Hello. I have been trying to run the multi task llama7b models with Bloke's llama 7b GPTQ(https://huggingface.co/TheBloke/Llama-2-7B-GPTQ) as the base.
```
def load_model(base_model, peft_model, …
-
#### Description
ModuleNotFoundError: No module named 'peft8'
#### To Reproduce
When I executed `bash run_NCE.sh`, it said that `ModuleNotFoundError: No module named 'peft8'`.
I checked `distill…
-
在 `colab` 环境,执行 **llama3** `lora` 微调报错:NotImplementedError: Cannot copy out of meta tensor; no data! 。微调代码如下:
```python
import pandas as pd
from modelscope import snapshot_download
import torch
f…
-
Hi there!
I'm exploring whether LISA is available on Hugging Face. I came across one link [here](https://huggingface.co/xinlai/LISA-13B-llama2-v0-explanatory), but unfortunately, there's no descrip…
-
I can't find a solution to this:
python generate.py --load_8bit --base_model 'decapoda-research/llama-7b-hf' --lora_weights 'tloen/alpaca-lora-7b'
===================================BUG REPORT==…
-
## ❓ Questions and Help
Hi, I recieved loss None when training model. Anyone can help?
Simple reproduct kaggle notebook [link](https://www.kaggle.com/code/liondude/notebook548442067d)
```
im…
-
[https://huggingface.co/upstage/solar-pro-preview-instruct](https://huggingface.co/upstage/solar-pro-preview-instruct)
Solar released a new 22b model, and this thing is crazy powerful. I was just won…
-
尝试autoeval , 评估chatglm2-6b,baichuan-7b 相同的错误如下:
1.chatglm2-6b:
python3 autoeval.py --model chatglm2-6b --lora_path ../../models/chatglm2-6b --eval_data all --device cuda:0
cuda:0
Lo…