Open cgbahk opened 3 years ago
cc @jinevening
For model compilation
For model quantization, right.
Inference speed was not the primary goal of luci-interpreter. So, there is a plenty of room for improvement, e.g., using faster kernel, parallizing jobs.
Recording multiple data in parallel would be a good option.
For model compilation, most time consumed on
record-minmax
(as I understand... is it right?), for some big production model about several minutes.As it just does inference on the model for many times, I guess 'how to split parallel job' is trivial.
So how about support option to run
record-minmax
in parallel? :smile: