Is your feature request related to a problem? Please describe.
Currently, when Triton Inference Server is running in --model-control-mode=explicit and a load_model request is sent from the client for a model with an error, the response results in a TimeoutError: timed out. This limited feedback provides insufficient information about the error, making it more difficult to debug and resolve issues with the model.
Describe the solution you'd like
I propose enhancing the Triton server's response mechanism in explicit mode. When an error occurs during the load_model process, instead of timeout, the server should return a detailed response containing Triton's log about the error. This would greatly aid in identifying and resolving issues with the model quickly and efficiently.
Describe alternatives you've considered
The only alternative now is to manually check Triton's logs, which is less convenient and time-consuming.
Additional context
This feature would make it easier and quicker to find and fix issues with models in the Triton Inference Server, leading to a smoother and more user-friendly model management experience.
Is your feature request related to a problem? Please describe. Currently, when Triton Inference Server is running in
--model-control-mode=explicit
and aload_model
request is sent from the client for a model with an error, the response results in aTimeoutError: timed out
. This limited feedback provides insufficient information about the error, making it more difficult to debug and resolve issues with the model.Describe the solution you'd like I propose enhancing the Triton server's response mechanism in explicit mode. When an error occurs during the
load_model
process, instead oftimeout
, the server should return a detailed response containing Triton's log about the error. This would greatly aid in identifying and resolving issues with the model quickly and efficiently.Describe alternatives you've considered The only alternative now is to manually check Triton's logs, which is less convenient and time-consuming.
Additional context This feature would make it easier and quicker to find and fix issues with models in the Triton Inference Server, leading to a smoother and more user-friendly model management experience.