Closed jphme closed 1 year ago
Hello, may I ask which GPU you are using?
It would seem that bitsandbytes are giving issues. Have you tried installing from source or downgrading one version?
This also kind of pushes the need of versioning #153
Hello, may I ask which GPU you are using?
It would seem that bitsandbytes are giving issues. Have you tried installing from source or downgrading one version?
This also kind of pushes the need of versioning #153
I was using a A100 40GB (via lambdalabs) and installed bitsandbytes via the requirements.txt:
accelerate==0.20.3
axolotl==0.1
bitsandbytes==0.39.1
peft==0.4.0.dev0
python_version: 3.9.17
After my experience I would also strongly support #153 , was about to give up...
@jpdus , may I ask if you have tried docker? I usually have best success with that.
I agree we need better versioning. It has been on the TODO for a long time..
Hello, did you manage to get this all setup and running? If so, can we close this issue?
Yes I managed to get everything up and running but spent the better part of 2 days on it ;-) .
Will close this, but would advice version pinning like proposed in #153 (or at least mentioning known "good" versions in the examples/tutorials; i created some requirements.txt lists and known "good" commits of the git installs for myself).
I used the setup script in the referenced issue (see here) install the correct pthon and pytorch version.
However, I got additional errors when running the examples, e.g.
accelerate launch scripts/finetune.py examples/openllama-3b/qlora.yml
or
accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml
results in
I´m not very experienced - are these axolotl issues or issues of the downstream libraries?
Edit:
After force-reinstalling tensorflow, protobuf and wandb, the examples are running again:
python -m pip install --upgrade --force-reinstall tensorflow "protobuf<3.21" wandb
Originally posted by @jpdus in https://github.com/OpenAccess-AI-Collective/axolotl/issues/242#issuecomment-1620549837