[TVM Discuss] [Questions] Performance of same op and workload in different model varies differently

2020-08-30 Thread nolan via TVM Discuss
There is no thread related ops. Besides, multi threads is faster than one threads. --- [Visit Topic](https://discuss.tvm.ai/t/performance-of-same-op-and-workload-in-different-model-varies-differently/7766/3) to respond. You are receiving this because you enabled mailing list mode. To un

[TVM Discuss] [Questions] Performance of same op and workload in different model varies differently

2020-08-30 Thread Chenfan via TVM Discuss
> However, when setting `OMP_NUM_THREADS=1` the model inference time is same, > seems it’s a problem with multiple threads. Will it be possible that there's any thread realated limitation in your pytorch script? --- [Visit Topic](https://discuss.tvm.ai/t/performance-of-same-op-and-work

[TVM Discuss] [Questions] Performance of same op and workload in different model varies differently

2020-08-30 Thread nolan via TVM Discuss
Compared two similar Bert models running on CPU with TVM, one is PyTorch model, the other is MXNet model. Due to the large performance difference, I did some profiling. The result shows the run time of the same operation(matmul) with same workload varies big. ENV: 1. TVM: build with MKL. 2.

[TVM Discuss] [Questions] Gotvm make error with cuda

2020-08-30 Thread kitkat via TVM Discuss
Hi, I have tried to implement gotvm as a GPU-accelerating deep learning runtime system on my edge computing architecture. When I make gotvm without cuda, there is no error. However, when I uncomment the cuda-related lines in tvm_runtime_pack.cc, the following errors are occurred. $ make

[TVM Discuss] [Questions] Unclear error from TE Inliner

2020-08-30 Thread masahi via TVM Discuss
Hi, I'm trying to compile the following trivial function: ``` fn (%p0: Tensor[(3), bool], Primitive=1) -> Tensor[(3), int32] { where(%p0, 1 /* ty=int32 */, -1 /* ty=int32 */) /* ty=Tensor[(3), int32] */ } ``` Note that the second and third arg to `where` is a scalar. Since this is not suppor