Open kimdodo97 opened 1 year ago
If possible, can you quantify what 'CPU usage is too high' is exactly? Like what % CPU usage, and what % GPU usage you get when running the model (and what you'd expect).
% CPU usage is very high In the Python model, it has a share of about 20%. Running it in C# it goes up to 50% How to solve?
How does the time to run the model compare vs Python? If it is faster, then the higher CPU usage could be a result of the GPU completing work faster and keeps the CPU busier.
See here for more information: https://onnxruntime.ai/docs/performance/tune-performance.html#why-is-my-model-running-slower-on-gpu-than-on-cpu
Currently, I am detecting real-time video through RSTP, and when I start video inference, it shows higher % CPU Usage compared to Python. When I output the inferred result in real time, the frame seems to drop
I'm having the same issue. The onnx model is the same as the original TF model on python, but significantly slower when running on C# (order of magnitude), also see the spike in CPU usage.
Describe the issue
I am using my YOLO model learned with pytorch by converting it to onnx I am inferring using onnxruntime-gpu in C#. GPU is used but CPU usage is too high Is there any way to lower the CPU usage?
Model name: YOLOv5s Model opset: 12
To reproduce
Urgency
No response
Platform
Windows
OS Version
window10
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.13.1
ONNX Runtime API
C#
Architecture
X64
Execution Provider
CUDA
Execution Provider Library Version
CUDA11.7
Model File
No response
Is this a quantized model?
No