Open Midi12 opened 2 weeks ago
I can take a look at it. Can you share the link to the failed compile job? Thanks!
I can take a look at it. Can you share the link to the failed compile job? Thanks!
Below are the related links, thanks.
https://app.aihub.qualcomm.com/jobs/jqpyy498p/ https://app.aihub.qualcomm.com/jobs/j0pxz7y85/
Hi @Midi12 , QCS6490
is RB-Gen2, which supports int8 only. I tried the same script but changed to the following options for compilation, it compiles successfully.
--target_runtime qnn_context_binary --quantize_full_type=int8 --quantize_io
To help quantization numerics, you would need to supply calibration_data
` to help with the quantization.
Hope this helps!
Hello @heydavid525 , yes this resolve my issue, thanks !
Hello, I am reopening this issue as I encountered the same error message with a larger model (Whisper medium from here https://github.com/openai/whisper/blob/main/whisper/__init__.py#L17).
Here is the links to related jobs https://app.aihub.qualcomm.com/jobs/j2p0w2425/ https://app.aihub.qualcomm.com/jobs/j7gj9z9ep/
One of two jobs is failing.
Describe the issue Please provide details relating to the issue you're hitting, if it is related to performance, accuracy or other model issues with bringing your own model to Qualcomm AI Hub, to deploy to IOT devices.
Hello, firstly thanks for the wonderful AiHub Platform :)
I am trying some example
To Reproduce Below the minimal example :
Using pre-trained MobileNet
torch_model = mobilenet_v2(pretrained=True) torch_model.eval()
Step 1: Trace model
input_shape = (1, 3, 224, 224) example_input = torch.rand(input_shape) traced_torch_model = torch.jit.trace(torch_model, example_input)
Step 2: Compile model
compile_job = hub.submit_compile_job( model=traced_torch_model, device=hub.Device("QCS6490 (Proxy)"), options="--target_runtime qnn_context_binary", input_specs=dict(image=input_shape), )
Expected behavior The model is compiled to a QNN context binary.
Compilation logs link1 link2
Host configuration:
Additional context Add any other context about the problem here.