triton-inference-server / server

The Triton Inference Server provides an optimized cloud and edge inferencing solution.
https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html
BSD 3-Clause "New" or "Revised" License
8.4k stars 1.49k forks source link

Build: Updating to allow passing DOCKER_GPU_ARGS at model generation #7566

Closed pvijayakrish closed 3 months ago

pvijayakrish commented 3 months ago

What does the PR do?

Updating the model generation script to allow passing DOCKER_GPU_ARGS directly.

Checklist

Commit Type:

Check the conventional commit type box here and add the label to the github PR.

Related PRs:

Where should the reviewer start?

Test plan:

Caveats:

Background

Related Issues: (use one of the action keywords Closes / Fixes / Resolves / Relates to)