NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html
Apache License 2.0
1.85k stars 309 forks source link

Parallel build with limited resource #981

Closed phu0ngng closed 3 months ago

phu0ngng commented 3 months ago

Description

Type of change

Checklist:

timmoon10 commented 3 months ago

We abandoned this PR based on the following chain of logic:

We're not the first to note how build isolation is poorly-suited for the ML ecosystem (see https://github.com/astral-sh/uv/issues/1715). We should keep this in mind for the future in case we need to modernize the build process and add a pyproject.toml. Users may want to preemptively run with pip install --no-build-isolation so that we don't break their build workflows.

Fow now, the much simpler approach is to modify our build process to handle either Ninja or make. See https://github.com/NVIDIA/TransformerEngine/pull/987.