NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.6k stars 255 forks source link

Expose `rotary_base` as an arg instead of hardcoding #944

Open sudhakarsingh27 opened 2 weeks ago

sudhakarsingh27 commented 2 weeks ago

Description

Expose rotary_base as an arg instead of hardcoding to 10000

Fixes # (issue) https://github.com/NVIDIA/TransformerEngine/issues/849

Type of change

sudhakarsingh27 commented 2 weeks ago

/te-ci pytorch