Closed simaotwx closed 1 month ago
Hey @simaotwx - sorry we missed this. May I ask what torch, huggingface and peft version you were running into this issue with? Are you still seeing it?
Hey @simaotwx - sorry we missed this. May I ask what torch, huggingface and peft version you were running into this issue with? Are you still seeing it?
It's okay, no worries.
Versions:
huggingface-hub 0.20.3
peft 0.8.2
sentence-transformers 2.3.1
torch 2.2.0
torchaudio 2.2.0
torchdata 0.7.1
torchinfo 1.8.0
torchmetrics 1.3.1
torchtext 0.17.0
torchvision 0.17.0
transformers 4.37.2
I am still seeing the same issue. I have since moved on and am not using ludwig anymore so this issue isn't relevant for me now but it still exists.
While looking at this issue I noticed that I copied the wrong config. I updated it to match what I just tested.
Hello, facing the same issue, any idea?
@simaotwx Did your example finally work? Thank you.
Describe the bug When trying to train (fine-tune) a Mistral model, there is an error when not using quantization.
To Reproduce Steps to reproduce the behavior:
Please provide code, yaml config file and a sample of data in order to entirely reproduce the issue. Issues that are not reproducible will be ignored.
Config file (model.yaml):
Command:
ludwig train --config model.yaml --dataset "ludwig://alpaca"
Experiment description:
User-specified config (with upgrades):
Expected behavior Training would start.
Screenshots
Environment (please complete the following information):
Additional context With 8-bit quantization it works: