Closed joyalbin closed 3 years ago
Hi sorry for the huge delay. TFLite working on 4bit quantization, but I doubt currently TFLite can support this with options. @daverim any ideas?
Hi, xhae@ is correct, we don't have a 4bit scheme in TFLite, however, it sounds like you want to convert your 4bit weights to 8bits with the range fixed to -7 to 7. We don't have a feature that can do this because the TFLite converter converts float weights to 8 bit regardless of the configured number of bits in the fake_quant, using only the min/max range.
TLDR: we don't support your use-case currently.
Hi facing an issue with TfLiteConverter.
Created a mode through QAT using TFMOT framework with 4bit training. While converting the model to tflite using TfLiteConverter, the model parameters still in INT8 data range
We need to convert the model with 4bit data range and its type in INT8.
Expected Result:
Actual Result:
Please help me how to resolve this.
Thanks Albin