isarsoft / yolov4-triton-tensorrt

This repository deploys YOLOv4 as an optimized TensorRT engine to Triton Inference Server
http://www.isarsoft.com
Other
277 stars 63 forks source link

Client side questions #14

Closed ontheway16 closed 3 years ago

ontheway16 commented 3 years ago

Hi, from documents, triton supports different versions of models, under /1/, 2, 3, etc. and newest one is considered /1/ folder.

How can we indicate which version of a model used in a specific inference request, via client.py ? I may prefer an older version sometimes. If its working this way.

philipp-schmidt commented 3 years ago

Hi,

yes the SDK supports passing the flag in most calls. Have a look e.g. here: https://github.com/triton-inference-server/server/blob/92099af2a1a2915747f5647d5acd35c0799e4730/src/clients/python/library/tritonclient/grpc/__init__.py#L949

The default value is model_version=""