-
Hello,
First of all, thank you for the fantastic work on the `openvino.genai` project. It's impressive how you've enabled inference for large language models and Stable Diffusion using the OpenVINO…
-
I've run the benchmark_genai.py for CPU, GPU, NPU on MTL U9, here is the logs:
(env_ov_genai) c:\AIGC\openvino\openvino.genai\samples\python\benchmark_genai>python benchmark_genai.py -m c:\AIGC\openv…
-
Dears,
I failed to run Llama-2-7b-chat-hf on NPU, please give me a hand.
1. I converted the mode by below command, and got two models,
a) optimum-cli export openvino --task text-generation -m Meta-…
-
I have downloaded LLAMA 3.2 1B Model from Hugging face with optimum-cli
optimum-cli export openvino --model meta-llama/Llama-3.2-1B-Instruct llama3.2-1b/1
Below are files downloaded
!…
-
**Describe the bug**
openVINO gave RuntimeError at Notebook:Convert Detectron2 Models to OpenVINO with Integrated GPU.
I changed openVINO version from 2023.1.0 to 2024.3.0, all gave RuntimeError.
…
-
It is my first time using audacity and I want to use the transcription features.
I have already installed ffmpeg and all the available language models with no issues during installation.
When I open…
-
I attempted to run benchmarks for the llama-3-8b-instruct and llama-3.1-8b-instruct models using both CPU and GPU, but the process failed. (I successfully tested the llama2-7b-chatbot model)
I f…
-
Trying to run chatQ&A with vLLM-openVINO and found vllm-openvino is not available as prebuilt image. Also having a helm chat for ChatQ&A with vLLM-openvino as model server will help to run on Xeon wit…
-
I'm curious to know how these techniques can be applied to Computer Vision models that need to be deployed on OpenVINO and TensorRT. I suspect that it will take a long to get a similar support on Open…
-
By using this model from Intel :
https://docs.openvino.ai/2024/omz_models_model_age_gender_recognition_retail_0013.html
I can't get good results (Or this model offers really good accuracy in the …