Open gkumbhat opened 1 year ago
Verify that we are able to fine-tune a model with multi-gpu sharding (fsdp) with bf16.
Provide detailed discussion here
Was this test covered by the finetuning testing @olson-ibm did? If not, is it definitely worth doing or not neccessarily?
Multi GPU support for finetuning is available, so this specific ticket should not be blocked.
Description
Verify that we are able to fine-tune a model with multi-gpu sharding (fsdp) with bf16.
Discussion
Provide detailed discussion here
Acceptance Criteria