triton-inference-server / server

The Triton Inference Server provides an optimized cloud and edge inferencing solution.
https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html
BSD 3-Clause "New" or "Revised" License
8.23k stars 1.47k forks source link

No 24.05-trtllm-python-py3 in NGC Repo #7277

Closed avianion closed 4 months ago

avianion commented 5 months ago

All other images like VLLM etc are available, but not this one. What gives?

24.05-py3-min 24.05-py3-sdk 24.05-pyt-python-py3 24.05-tf2-python-py3 24.05-vllm-python-py3 24.05-py3 24.05-py3-igpu-min 24.05-py3-igpu-sdk 24.05-py3-igpu

But no 24.05-trtllm-python-py3 ??

krishung5 commented 4 months ago

Hi @avianion, the TRT-LLM container is delayed. We will publish it as soon as it's available. Thank you.

krishung5 commented 4 months ago

The TRT-LLM image is out. Closing the issue.