Open navyverma opened 3 months ago
That's something you would want to debug, or give us the model if possible. Otherwise, there is nothing to go with.
@yuslepukhin I am asking whether FP16 implementation is correct or not?
@navyverma,
CPU provider does not support FP16 for most operators. So it does not make sense to run float16 in CPU. Even float16 model can run (internally ORT will add fp16 <-> fp32 conversions), float32 model possibly runs faster than float16 model in CPU.
If you want to get benefit of fp16 model, try CUDA provider on a GPU instead.
Also, try onnxruntime 1.19 which might have fixed some bugs.
@tianleiwu Thanks for your response. I will try latest onnxruntime.
That's something you would want to debug, or give us the model if possible. Otherwise, there is nothing to go with.
Have you guys completely missed the fact that if the model input is float16 it doesn't work at all? float16 mode output is completely wrong, mainly because of the data preprocessing stage, but it works fine in python.
Describe the issue
I have a detector with FP16 and FP32 weights(onnx). Below is the code for FP32 which gives the correct detections when inferencing on FP32 weights.
Below is the code for FP16 which gives the correct grabage detections when inferencing on FP16 weights.
What could be the issue in FP16 inferencing?
To reproduce
Use above process_image16 method to reproduce the issue.
Urgency
No response
Platform
Linux
OS Version
20.04
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.17.3
ONNX Runtime API
C++
Architecture
X64
Execution Provider
Default CPU
Execution Provider Library Version
No response