Open ivan94fi opened 2 years ago
I'm having a similar issue, don't know if it is the FP16 overflow issue or InstanceNorm issue.
Hi,
I'm having a similar issue trying to convert to FP16. The conversion was working fine with previous versions of pytorch and this started to happen when I upgraded torch from 1.8 to 1.10.
So one quick fix for you would be to downgrade to earlier version of pytorch and it might solve your problem. It is not a long term solution though because we need to upgrade at some point ! We will need to find the root of this problem aniway..
For the moment I'm trying to read the changelog to see if there is something critical that changed.
Thanks for these information, however downgrading pythorch is not an option for us.
Any updates from the developers on this?
Thank you
I have a problem with instance normalization, the model outputs diverge substantially when using the tensorrt model with float16 precision.
Versions:
Example code:
This is the output I get from this script:
The errors are too big in float16 mode, especially considering this is just one instance normalization and there are many in my model, so errors are propagated and become larger.
This may be caused by the fact that instance norm uses float16 precision and this causes numerical errors.