Open hayyaw opened 1 month ago
can not convert tinyllama to full int8 quantized tflite model
The compute platform only support int8 datatype, request for tflite full int8 static quantization with calibration dataset.
No response
Noted.
Description of the bug:
can not convert tinyllama to full int8 quantized tflite model
Actual vs expected behavior:
The compute platform only support int8 datatype, request for tflite full int8 static quantization with calibration dataset.
Any other information you'd like to share?
No response