Open danielholanda opened 1 year ago
It's a segfault. The easiest way to repro is to go into devices.py https://github.com/groq/mlagility/blob/54a774cbcbded984c213798acdd2aea43f05b7ee/src/mlagility/api/devices.py#L609
and print out the command that is being used in the subprocess. From there you can run the command yourself to see the segfault happen, or open it up in a python debugger.
This is due to the FP16 converter issue. I have tested a bunch of these cases and the base
model passes but the FP16 version that we use for benchmarking fails to run.
I would love to add a FP32 version of the benchmark and have that as a toggle on the dashboard.
@ramkrishna2910 I guess the failures are not user error then? In the sense that ORT does not support a variety of valid fp16 ONNX files?
Yes, these are not user errors. This is due to some issues in the FP16 converter. I have created an issue upstream as well on this. @danielholanda and I spoke about adding the FP32 numbers, but we were worried that it will make the dashboard too chaotic and comparing FP32 CPU with FP16 GPU may not be right.
Issue:
OnnxRuntime seems to be unable to execute half of our benchmarks. Therefore, it may not be appropriate as a baseline.
For example, simply running
benchit models/torch_hub/alexnet.py --device x86
results in the following error:Task: