NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.85k stars 309 forks source link

Add a CP implementation variant with KV all-gather. #1060

Closed xrennvidia closed 1 month ago

xrennvidia commented 2 months ago

Description

This is a CP implementation variant with KV all-gather. Currently, it can support:

Will add more functionality support later.

The KV all-gather communication is exposed, but the overheads should be small with GQA/MQA.

Type of change

Changes

Please list the changes introduced in this PR:

Checklist:

xrennvidia commented 1 month ago

/te-ci

xrennvidia commented 1 month ago

/te-ci pytorch

xrennvidia commented 1 month ago

/te-ci pytorch

xrennvidia commented 1 month ago

/te-ci pytorch

cyanguwa commented 1 month ago

/te-ci pytorch