Thanks you for replying!

using mkl_verbose=1, i found that tvm_num_threads do not affect threads used by 
mkl.

so i used mkl_num_threads and resolved problems(fluctuation, slow)

with and without -libs=mkl, the inference time is measured approximately the 
same.

While searching for the reason, I found out that tvm uses mkl to optimize the 
only dense layer.
I also found that auto-tvm can tune the dense layer.

As a result, if not use mkl, the tvm default tuning option applies to all 
layers.
If use it,  mkl would be applied a dense layer.
is this right?


If I'm right,
auto-tvm (default) 
auto-tvm + mkl(only dense)

these two cases show similar performance, and can i say that the schedule 
primitives of tvm show as much performance as mkl?





---
[Visit 
Topic](https://discuss.tvm.apache.org/t/inference-time-fluctuation/9213/5) to 
respond.

You are receiving this because you enabled mailing list mode.

To unsubscribe from these emails, [click 
here](https://discuss.tvm.apache.org/email/unsubscribe/0f5f6800350c95e124e53b5e78918d59e9604b363f696506326f362768be693b).

Reply via email to