triton-inference-server / model_analyzer

Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.
Apache License 2.0
435 stars 75 forks source link
deep-learning gpu inference performance-analysis

License

Triton Model Analyzer

[!Warning]

LATEST RELEASE

You are currently on the main branch which tracks under-development progress towards the next release.
The latest release of the Triton Model Analyzer is 1.46.0 and is available on branch r24.11.

Triton Model Analyzer is a CLI tool which can help you find a more optimal configuration, on a given piece of hardware, for single, multiple, ensemble, or BLS models running on a Triton Inference Server. Model Analyzer will also generate reports to help you better understand the trade-offs of the different configurations along with their compute and memory requirements.

Features

Search Modes

Model Types

Other Features

Examples and Tutorials

Single Model

See the Single Model Quick Start for a guide on how to use Model Analyzer to profile, analyze and report on a simple PyTorch model.

Multi Model

See the Multi-model Quick Start for a guide on how to use Model Analyzer to profile, analyze and report on two models running concurrently on the same GPU.

Ensemble Model

See the Ensemble Model Quick Start for a guide on how to use Model Analyzer to profile, analyze and report on a simple Ensemble model.

BLS Model

See the BLS Model Quick Start for a guide on how to use Model Analyzer to profile, analyze and report on a simple BLS model.

Documentation

Terminology

Below are definitions of some commonly used terms in Model Analyzer:

Reporting problems, asking questions

We appreciate any feedback, questions or bug reporting regarding this project. When help with code is needed, follow the process outlined in the Stack Overflow (https://stackoverflow.com/help/mcve) document. Ensure posted examples are: