Thanks for the great tool for model compression research.
As it is stated that you guys are planning to add PTQ or capability to export the quantized model to ONNX, would be greatly interested in knowing if that's yet in the pipeline, or it's better to consider PyTorch internal quantization capabilities or using TensorRT for the same.
Hi Team,
Thanks for the great tool for model compression research.
As it is stated that you guys are planning to add PTQ or capability to export the quantized model to ONNX, would be greatly interested in knowing if that's yet in the pipeline, or it's better to consider PyTorch internal quantization capabilities or using TensorRT for the same.
Thanks!