NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.6k stars 255 forks source link

Why is the result of context-parallel DotProductAttention influenced by the random seed? #961

Open LitPrice opened 1 week ago

LitPrice commented 1 week ago

Hi! When I want to replace the regular attention calculation with context-parallel DotProductAttention, I find that the results of DotProductAttention are influenced by different random seeds, and the outputs are not completely aligned. How can I resolve this situation?