Closed ladanisavan closed 2 months ago
Hi @ladanisavan please open a issue on the onnxruntime github https://github.com/microsoft/onnxruntime and this issue is related to the instructions provided at https://onnxruntime.ai/docs/genai/howto/build-model.html
Please open a discussion issue on ONNX repo or the actual model repo.
This issue is for a: (mark with an
x
)Hi there,
I'm seeking guidance on exporting a custom fine-tuned Phi-3 Vision model to ONNX. I've followed the ONNX build model guide from this link.
The build command I used was:
python3 -m onnxruntime_genai.models.builder -i ep_2_grad_32_lr_3e-5/ -o onnx_output/ -p int4 -e cuda --extra_options int4_block_size=32 int4_accuracy_level=4
The build process was successful and generated the following files:
However, the number of files generated doesn't match the file count in the official HF repo for ONNX
microsoft/Phi-3-vision-128k-instruct-onnx-cuda
Files highlighted in red below are missing:
Additionally, while loading the model using ONNX Runtime, the following error occurs:
OrtException: Load model from onnx_output failed: Protobuf parsing failed.
I have also noticed that sections for "embedding" and "vision" are missing from the
genai_config.json
Can someone help me identify if I'm missing anything? Thanks