when I auto tune a modified mobilefacenet, I encountered the following error
if I use mxnet to inference, the time is 41ms. when I use apply_history_best(“model.log”) the time is 360ms.
if I just use relay.build without autotvm, the time is 390ms. it is so weird.
Did you set appropriate llvm target?
yes, target = tvm.target.create(“llvm -mcpu=core-avx2”)
if i use gpu, target = “cuda”, the time is 5ms.
when i build with opt_level=1 or opt_level=2, the time is 130ms. opt_level=3 is 390ms
this is the model, you can check it if you have time. or can you tell me how to find out the problem.
You can use debug_runtime to see whether conv2d execution takes majority of the time.