Open umadevimcw opened 1 month ago
@ttmtrajkovic I have discussed with @umadevimcw offline, and the issue above is that exp2 implementation on device does not output value: 1.1977575e-21
, which is re-presentable by float16b, but pytorch with dataformat float16b does represent it, this causes precision failures downstream.
@umadevimcw @eyonland please let us know what is the priority for this issue, if it is only failing unit tests, or failing on models due to the downstream precision issue.
@ttmtrajkovic can re-assign appropriately.
This is a P1 priority. I'm not aware of any models failing on this at the moment. The related issue is #8634
Describe the bug PCC is dropping due to precision loss for lodaddexp2 functions. While debugging observed that exp2 of certain inputs are zeros whereas in Torch we are getting values at the precision level which results in PCC drop.
To Reproduce Steps to reproduce the behavior:
Copy Paste this code to get the precision loss. In this code the input values are fixed for debugging purposes which show cases the precision loss
Expected behavior
Screenshots
Please complete the following environment information:
Additional context TT's exp2 ops internally depends on exp op