I’m having some problems when deploying this model on x86 architecture using relay.
opt_level=3 performance get worse by almost 3 times compared to using
Any clue why does this happen?
I’ve also tuned convolution on x86 using
autotvm, but inference is still quite slow. Is there any other way I can further speed up the inference?