The FSDP example launches Llama2 7b fully sharded data parallel distributed training on the huggingface c4 dataset.
The smpv2 example launches SageMaker ModelParallel Llama2 7b distributed training on HyperPod EKS.
By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.
Issue #, if available:
Description of changes:
The FSDP example launches Llama2 7b fully sharded data parallel distributed training on the huggingface c4 dataset. The smpv2 example launches SageMaker ModelParallel Llama2 7b distributed training on HyperPod EKS.
By submitting this pull request, I confirm that you can use, modify, copy, and redistribute this contribution, under the terms of your choice.