Closed rag1998 closed 5 years ago
To the best of our knowledge, TensorFlow and TensorFlow Lite does not support directly inference using models with non-structured sparsity and stored in CSR/CSC format. Therefore, we do not provide such model conversion tool. A possible solution to manually create a model loader so that the model can be stored in CSR/CSC format on disk and de-compressed to memory during inference. But this can only reduce the disk space consumption, which seems not so important in many applications.
We are trying to evaluate Weight Sparsification Learner for some compression needs, using Optimal Prune Ratio Protocol we were able to achieve good accuracy but are having a bit of a hard time figuring out a converter to use to get .pb and .tflite files.
Is there one built already or do we have to create one?