Open 12sf12 opened 3 months ago
@12sf12 how did you export this model? Can you share the trtexec command?
Also, can you share complete config.pbtxt for this model? If you are passing max_batch_size as 0, the batch size has to be mentioned in the tensor definition.
Description I'm having a strange issue with integrating a tensorrt model into Triton. When I retrieve the model configuration, I see that the max_batch_size is being considered as the number of channels for a
3*H*W
image input. For example, Triton returns this configuration for aC*H*W
image:max_batch_size=C=3
and dims=H*W
. I want to note that the model works fine in the Python environment and I have already received correct results from it.Triton Information
Are you using the Triton container or did you build it yourself? Just use it with no modifications
To Reproduce
curl localhost:8000/v2/models/txspot/config
For the above example, the dims should have been
3*1152*2048
and max_batch_size=1, while Triton returnedmax_batch_size=3
and dims=1152*2048
Describe the models (framework, inputs, outputs), ideally include the model configuration file (if using an ensemble include the model configuration file for that as well).
the config.pbtxt is:
Expected behavior The dimensions should have been
C*H*W
, but Triton considers the number of channels (C) as the max_batch_size and the dimensions asH*W
. So the max_batch_size is 3, which is equal to C.