autodeployai / ai-serving

Serving AI/ML models in the open standard formats PMML and ONNX with both HTTP (REST API) and gRPC endpoints
Apache License 2.0
148 stars 31 forks source link

AI-Serving not utilizing GPU #3

Closed ajertec closed 4 years ago

ajertec commented 4 years ago

Hello, I got AI Serving Server up and running:

However, when trying out mnist example (AIServingMnistOnnxModel.ipynb notebook), or modifying this example and deploying my custom models with gpu optimizations (https://pypi.org/project/onnxruntime-tools/) server is not utilizing the GPU, only CPU.

My models metadata: (this is actually TinyBERT transformer model) image

Can you provide full example of using the AI-Serving with GPU utilization?

Thank you.

scorebot commented 4 years ago

@ajertec Thanks for your issue. It's a defect of the AI-Serving, we will fix it as soon as possible, and let you know once it's done.

scorebot commented 4 years ago

@ajertec We have fixed the issue of GPU. Please, try the image autodeployai/ai-serving:0.9.1-cuda that if resolve your issue.

ajertec commented 4 years ago

@scorebot Thank you, now GPU is being utilized.

scorebot commented 4 years ago

Cool. I close it now.