Closed cyrusbehr closed 5 years ago
Please open a new thread on https://discuss.tvm.ai/ for related discussions
Hi @tqchen I made a post on https://discuss.tvm.ai/ but have been waiting for approval of my post from a staff member for over 2 days. Can you kindly approve my post. I am keen to get an answer.
Have your performance improved? I encounter the same problem as yours. @cyrusbehr
I am trying to convert an MXNET model to TVM in order to improve the inference speed. I am able to convert it successfully, however I do not experience the improvements in speed which are advertised on this page
I have followed the tutorial here, but I will go through the steps I took.
I first downloaded the Insightface model
LResNet100E-IR,ArcFace@ms1m-refine-v2
which can be found here. Note that I am using the same model from the TVM benchmark.Next, I use the following python script to convert the model to the TVM compatible models Note that when I run the command
llc --version
I get the following output (which is why I set the target toskylake
)Python conversion script
When I run the script, I get the following warning messages:
but it ultimately exports the models successfully.
Next, I import the the converted models and
deploy_lib.so
into my C++ project. I am using the following code. The majority of the code is taken from the example on this pageNote that the images I am provided are pre-aligned and cropped to 112x112.
On average, the inference takes 360ms, which is roughly the same time it takes to perform inference using MXNET (C++, MKLDNN). I was expecting to see a significant decrease in inference time.
I am not sure if the issue has to do with the warnings during the conversion? I followed the conversion tutorial exactly and the tutorial did not mention needing to fine tune the model or anything.
Here is the output of
cat /proc/cpuinfo
to understand what hardward I am running the benchmark on: