Open cab1211 opened 1 year ago
the model precision drop from 90% to 51% when quantize from float32 to int8, but didn't drop with the int16.
the model precision drop from 90% to 51% when quantize from float32 to int8, but didn't drop with the int16.