Open cahuja1992 opened 1 year ago
Same issue for me as well.
Am just doing with single gpu for now but running into OOM issue - https://github.com/rmihaylov/falcontune/issues/19
I have the same issue , device_map=auto does not work for training. I guess we should copy tokenizer to each and all devices ? Why does device_map=auto can not handle this ?
Same here. While training on multi-GPU I get the following error:
ValueError: DistributedDataParallel device_ids and output_device arguments only work with single-device/multiple-device GPU modules or CPU modules, but got device_ids [0], output_device 0, and module parameters {device(type='cuda', index=0), device(type='cuda', index=1)}.
It seems it doesn't rely on Accelerate framework.
An easy quick win would be to rely on the new HuggingFace SFT Trainer instead: https://huggingface.co/docs/trl/sft_trainer
Does fine-tuning support the multi-GPU training?
When trying to fine-tune with multiple GPUs, got the following error.