NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.85k stars 309 forks source link

Parallel build with limited resource #987

Closed phu0ngng closed 2 months ago

phu0ngng commented 3 months ago

Description

This PR adds an option for users to specify the maximum number of involved threads for the compilation of cpp codes by NVTE_MAX_BUILD_JOBS or MAX_JOBS env vars.

It works with both Ninja and Make.

Make now build with parallel by default.

Type of change

Checklist:

timmoon10 commented 3 months ago

/te-ci