🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
This PR changes the zero_optimization.stage3_prefetch_bucket_size config value to be an integer. With the impending Pydantic v2 update in DeepSpeed, we run into validation errors if the value is a float. The documentation states that this value should be an integer.
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
This PR changes the
zero_optimization.stage3_prefetch_bucket_size
config value to be an integer. With the impending Pydantic v2 update in DeepSpeed, we run into validation errors if the value is a float. The documentation states that this value should be an integer.cc @muellerzr