NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.6k stars 255 forks source link

[PyTorch] Fix invalid import in test for context parallelism #968

Closed timmoon10 closed 1 week ago

timmoon10 commented 1 week ago

Description

The _cudnn_version function was removed from the fused attention tests in https://github.com/NVIDIA/TransformerEngine/pull/889, but not from the corresponding tests for context parallelism. This has already been fixed in the main branch with https://github.com/NVIDIA/TransformerEngine/pull/927, but not in the 1.8 release.

Type of change

Changes

Checklist: