NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.61k stars 256 forks source link

[PyTorch] Expose `multi_tensor_*` kernels #907

Closed yaox12 closed 3 weeks ago

yaox12 commented 3 weeks ago

Description

Expose multi_tensor_* kernels under transformer_engine.pytorch.optimizers to avoid importing transformer_engine_torch outside of TE.

Type of change

Checklist:

yaox12 commented 3 weeks ago

Renaming transformer_engine_extensions to transformer_engine_torch makes me realize that we should avoid importing it outside of TE, so I expose these kernels to let MCore (or other framework) use them. cc @timmoon10

timmoon10 commented 3 weeks ago

/te-ci pytorch

yaox12 commented 3 weeks ago

@timmoon10 Please re-run the CI.

timmoon10 commented 3 weeks ago

/te-ci pytorch