NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
2k stars 333 forks source link

[C] Separating cudnn common utils from fused_attn #1314

Closed phu0ngng closed 2 weeks ago

phu0ngng commented 3 weeks ago

Description

In this PR, the common cudnn helpers are moved out of fused attention so that they can be used in other kernels, i.e. the coming normalization.

Fixes # (issue)

Type of change

Checklist:

phu0ngng commented 3 weeks ago

/te-ci pytorch

phu0ngng commented 3 weeks ago

/te-ci pytorch