Closed hcho3 closed 4 months ago
Use TRITONBACKEND_ModelInstanceReportBatchStatistics to report statistics about batching.
TRITONBACKEND_ModelInstanceReportBatchStatistics
Example: https://github.com/triton-inference-server/tensorflow_backend/blob/515466c5760612aad0458d5e4f832c4114163271/src/tensorflow.cc#L2548-L2553
Already implemented in Rapids-Triton.
Use
TRITONBACKEND_ModelInstanceReportBatchStatistics
to report statistics about batching.Example: https://github.com/triton-inference-server/tensorflow_backend/blob/515466c5760612aad0458d5e4f832c4114163271/src/tensorflow.cc#L2548-L2553