NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.61k stars 256 forks source link

Use correct FP8 group in multi-GPU docs #852

Closed timmoon10 closed 1 month ago

timmoon10 commented 1 month ago

Description

The process group for FP8 amax reductions should be the tensor-parallel group, but we use the data-parallel group in the docs for multi-GPU training. See https://github.com/NVIDIA/TransformerEngine/issues/656#issuecomment-1972000820.

Fixes https://github.com/NVIDIA/TransformerEngine/issues/656.

Type of change

Changes

Checklist: