Closed aiqiangbrave closed 1 year ago
Hi @aiqiangbrave, Here is the answer of the first question. The const data stored in the provided model "YOLOX_0_int.xmodel" are still in floating-point format. The data will be converted to XINT8, and the output size will be approximately 25% of the original size after compiling. I have tried compiling the model using zcu104, the output xmodel size is 27MB.
For the second question, please check the model architecture to see if there are any modifications comparing with the official version. If your quantized.xmodel is different from the official version, there may exist some unencountered issues. I have used the official version and compiling it on zcu102, zcu104, and kv260 and doesn't reproduce your problem.
Could you please provide your quantized.xmodel, your architecture information and your vitis-ai version, so that we can reproduce your scenario.
Closing since no activity for more than 1 month, please reopen if you still have any questions, thanks
Could you help me analyze the size of the model? I remember that the xmodel model after yolov4 QAT is about 7M, but the model after yolox QAT is only twice shrunk (about 100M), is this normal? Why is the quantified model still so large?
Thank you so much~~~
now,I ignored the previous question for a moment. I used the run_train.sh trained float model to generate the xmodel file, and although it was successful, I compiled it with the following error: Instruction format: vai_c_xir -x /PATH/TO/quantized.xmodel -a /PATH/TO/arch.json -o /OUTPUTPATH -n netname error: [UNILOG][FATAL][XCOM_PM_FAIL][The compiler occurs an error when generating instructions, please contact us.]