Open ek9852 opened 4 years ago
A few initial questions:
I don't think the post training quantization tool supports GPU but I'm not the expert. I'll let @suharshs follow from here.
For "test set of 2336 on our model", does it mean 2336 images are used as representative dataset? Yes
Do you know how much time does it take to invoke the model? 0.3 sec on a Google Coral edge tpu. It should be much faster on my TITAN X nvidia gpu, But post-quantization does not use GPU currently.
TensorFlow Lite doesn't currently support non-mobile GPU kernels, and the post-training quantization tool is specific to TensorFlow Lite at the moment. As we work to unify TensorFlow and TensorFlow Lite we will keep this in mind. I will keep this issue open to give you updates as they come.
Thanks!
System information
Motivation During post quantization , the GPU is idle (confirmed via nvidia-smi ), i.e. the post quantization is not using GPU to speed things up. It is very slow. It takes > 60 min to run on a server grade xeon (for test set of 2336 on our model):
Describe the feature post quantization should utilize GPU to speed things up.