Closed dhernandez0 closed 1 month ago
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 83.55%. Comparing base (
c9e737f
) to head (f8293de
). Report is 10 commits behind head on develop.
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
- There's a discrepancy between the
populate()
calls and the command-line setup which could be confusing. We might want to resolve that, and I don't have a strong opinion on which direction to go with it. (That is, if we want the populate()s to take an extra set, we'd hardcode F32 and F64 into the type converter). But that's up to you
I understand you mean calling addConversion() for f64 and f32 (and ShapedTypes of f64 and f32)? That seems like duplicating the code to me. I'm not sure how to do this cleanly, another way would be to pass MLIRContext to populateExtendToSupportedTypesConversionTarget() and there just copy the extraSupported vector and add f64 and f32. I'm not even sure why those populate functions are public, they seem internal stages of the pass?
In this PR we add f16 math library calls.
TODO:
~test MIGraphX model (accuracy)~
~check if there's any performance improvement~ roughly same performance for bert_base_cased_1_fp16_gpu, distilgpt2_1_fp16_gpu and resnet50-v1-7 (--fp16)
upstream PR?
closes ticket: https://github.com/ROCm/rocMLIR-internal/issues/1439