-
### Describe the issue
While installing onnxruntime-openvino via pip, the openvino package is not installed by default. One doesn't know:
1. they must install openvino separetely
2. which version…
-
What could be the problem, please help..
Thank you
-
During model inference, model weight is frozen and won't change between iterations. CPU prefers special weight layout to accelerate the execution, then we need to prepack the model weight before model…
-
### Search before asking
- [X] I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar bug report.
### YOLOv8 Component
_No response_
### Bug
…
-
-
### Describe the issue
onnxruntime + openvino need double memory compared with openvino-only
I guess the onnx model and opnevino model are in the memory in the same time.
my model size: 330M
onnx…
-
### 🚀 Feature request
Quantization is a widely used technique to accelerate models, particularly when using the [torch.compile](https://pytorch.org/tutorials/intermediate/torch_compile_tutorial.htm…
-
Greetings, @cipher982!
Currently we are working on the Openvino inference framework, and such benchmarks are critical to understand gaps and differences between our framework and Transformers/ TGI …
-
### Package name
openvino
### Package homepage
https://www.intel.com/content/www/us/en/developer/tools/openvino-toolkit/overview.html
### Description
OpenVINO is an open-source toolkit for optimi…
-
### OpenVINO Version
2024.1.0
### Operating System
Windows System
### Device used for inference
NPU
### Framework
None
### Model used
Matmul
### Issue description
I'd …