Closed alexkehoe closed 1 year ago
Hi, please read the quickstart guide to see details for the Tritonserver.
I believe there are users that have used yolo models with Triton. We require the data to be sent from Triton clients. Please read up about the Triton Architecture.
https://github.com/triton-inference-server/client#triton-client-libraries-and-examples has some more examples you can refer to
@jbkyang-nvi all those examples are using image files or image folders. Are there any examples with video streams?
Similar question: https://github.com/triton-inference-server/server/issues/4487
https://developer.nvidia.com/deepstream-sdk to learn more about deepstream
Closing issue due to lack of activity. Please re-open the issue if you would like to follow up with this issue
Hi we have a yolov model that we've converted to tensorRT for inference on the edge. We'd like to port this model to the cloud so we can process +100 streams. Can triton be used to run such a model in the cloud? Ie we would send an RTSP or WebRTC video stream to triton and then send just the bounding box results back to our local server for display to the user? FPS would be say 10hz
If Triton is not appropriate for this use case, can you recommend another way of trying to scale deployment of a yolo model for +100 video streams?