NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.85k stars 309 forks source link

[C] Allow bias support for sm80/86/89 for cuDNN 9+ #863

Closed cyanguwa closed 4 months ago

cyanguwa commented 4 months ago

Description

This PR opens up the restriction on post_scale_bias support for sm80/86/89 when cuDNN 9.0+ is installed.

Fixes QA tests for test_cp_with_fused_attention on sm80/86/89.

Type of change

Changes

Checklist:

timmoon10 commented 4 months ago

/te-ci