Open waittim opened 3 years ago
Did you ever resolve this? I'm facing the same problem, even had the same idea to try quantizing
I just moved to another approach.
I'm having the same problem at the moment. What approach did you end up using instead?
DynamicQuantizeLinear:
Resize:
The onnx model is quantized by onnxruntime.quantization.quantize_qat.
The original onnx model is converted from PyTorch model(.cfg&.pt). Both the original onnx model and the quantized onnx model can work in the Python environment. When I'm using the original onnx model, it shows that 'Uncaught (in promise) TypeError: int64 is not supported'. Therefore, I quantized it to int8 by onnxruntime.quantization. However, it still doesn't work. you can find the model I used in this folder.