openvinotoolkit / openvino

OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
https://docs.openvino.ai
Apache License 2.0
6.83k stars 2.18k forks source link

[Performance]: inference takes too long on simple tasks #26264

Open xueyingxin opened 2 weeks ago

xueyingxin commented 2 weeks ago

OpenVINO Version

2021.2.1.0

Operating System

Windows System

Device used for inference

CPU

OpenVINO installation

Build from source

Programming Language

C++

Hardware Architecture

x86 (64 bits)

Model used

ssd

Model quantization

Yes

Target Platform

No response

Performance issue description

Normally a face detection task takes about 1ms on GPU and <5ms on CPU. But sometimes, the inference time increased to >100ms on CPU. The time computing only includes the api: InferRequest.infer(). Is this an known issue?

Step-by-step reproduction

No response

Issue submission checklist

wenjiew commented 2 weeks ago

Hi, thanks for reaching out. Do you see the same issue in recent version of OpenVINO like 2024.3? If so, please let us know a few key information on the model used, OS, and the CPU platform. Thanks!