Closed lilhoser closed 4 weeks ago
Hi @lilhoser, which version of ONNX Runtime GenAI are you running? And which version of the example? I cannot reproduce your issue
@natke thanks for the ping. I can't seem to repro this either. Perhaps I had mismatched runtimes with model type. Will close for now.
Describe the bug
The model will load but inference fails with a runtime error.
To Reproduce Steps to reproduce the behavior:
https://huggingface.co/microsoft/Phi-3-mini-128k-instruct-onnx
Phi-3-mini-128k-instruct-onnx\cpu_and_mobile\cpu-int4-rtn-block-32-acc-level-4
Expected behavior
It should support phi 3?
Screenshots
Desktop (please complete the following information):