Detection models now can be exported to TRT engine with batch size > 1 - inference code doesn't support it yet, though now they could be used in Triton Inference Server without issues.
Is there any plan for this? Or how can I implement batch inference myself?
Thank you for excllent work.
Is there any plan for this? Or how can I implement batch inference myself?