Open scampion opened 1 year ago
After recompiled bitsandbytes
from source with a compliant version of CUDA 11.7 supported by torch.
The issue is still there
My mistake, the example.py
doesn't support multi GPUs.
WIP
It's complicated. This fork got rid of many things required for multi-GPU usage. One way to restore that would be to create adapted versions of the model parallel layers in fairscale (https://github.com/facebookresearch/fairscale/blob/main/fairscale/nn/model_parallel/layers.py) that use bitsandbytes.
For the moment, I can't run the 65B model with 4 GPUs and a total of 96GB.
I investigate,
bitsandbytes was compiled without GPU support. 8-bit optimizers and GPU quantization are unavailable
are a first idea ...