NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.61k stars 256 forks source link

[JAX] Fixed the shape miss-matching issue in MLP. #859

Closed mingxu1067 closed 1 month ago

mingxu1067 commented 1 month ago

Description

Please include a brief summary of the changes, relevant motivation and context. This PR is to fix shaping miss-matching issues inbetween activations and its corresponding sharding contraints in MLP/FF1 when using unfused activation functions.

Fixes # (issue)

Type of change

Changes

Please list the changes introduced in this PR:

Checklist:

mingxu1067 commented 1 month ago

/te-ci jax

mingxu1067 commented 1 month ago

/te-ci jax