Is your feature request related to a problem? Please describe.
OpenVINO backend currently supports inference only on CPU devices using OpenVINO CPU plugin.
Describe the solution you'd like
I would like to perform inference on VPU devices (Myriad/Myriad X accelerators) using the OpenVino backend through Triton server by specifying the target device through the model configuration like:
Is your feature request related to a problem? Please describe. OpenVINO backend currently supports inference only on CPU devices using OpenVINO CPU plugin.
Describe the solution you'd like I would like to perform inference on VPU devices (Myriad/Myriad X accelerators) using the OpenVino backend through Triton server by specifying the target device through the model configuration like:
Describe alternatives you've considered Running models through OpenVINO directly or by switching to OpenVINO Model Server