[Apache TVM Discuss] [Questions] [op testing] single op ir testing batch matmul

2022-02-09 Thread chenugray via Apache TVM Discuss
[[BUG]fix batch matmul not set attrs_type_key, when using tvm.parse.parse_expr will raise error.](https://github.com/apache/tvm/pull/10209) --- [Visit Topic](https://discuss.tvm.apache.org/t/op-testing-single-op-ir-testing-batch-matmul/12049/4) to respond. You are receiving this because

[Apache TVM Discuss] [Questions] Tvmc tune extra options for OpenCL

2022-02-09 Thread p via Apache TVM Discuss
Hello, I have trouble with runing autotuning using OpenCL and RPC. When I call tvmc tune command with `--help` command I'm getting a lot of options, e.g.: ``` target opencl: --target-opencl-max_num_threads TARGET_OPENCL_MAX_NUM_THREADS target opencl max_num_threads -

[Apache TVM Discuss] [Questions] Any way to extract tir function from relay?

2022-02-09 Thread Jack Kosaian via Apache TVM Discuss
I'm also interested in this. @chenugray , did you find a way to get the TIR from a Relay op? Thanks! --- [Visit Topic](https://discuss.tvm.apache.org/t/any-way-to-extract-tir-function-from-relay/11907/2) to respond. You are receiving this because you enabled mailing list mode. To unsub

[Apache TVM Discuss] [Questions] [Pytorch] The inference results of tvm and pytorch are inconsistent

2022-02-09 Thread masahi via Apache TVM Discuss
Interesting, does PyTorch do something like that? It's not obvious to me if we can do this without concern. Would this change make the output of every adaptive avg pool different? What about normal avg pooling? --- [Visit Topic](https://discuss.tvm.apache.org/t/pytorch-the-inference-resul

[Apache TVM Discuss] [Questions] [op testing] single op ir testing batch matmul

2022-02-09 Thread masahi via Apache TVM Discuss
Good catch, can you send a PR to fix it? --- [Visit Topic](https://discuss.tvm.apache.org/t/op-testing-single-op-ir-testing-batch-matmul/12049/3) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click here](https://discuss.tvm.

[Apache TVM Discuss] [Questions] [op testing] single op ir testing batch matmul

2022-02-09 Thread chenugray via Apache TVM Discuss
ok, the op Registry do have some problem. ![image|690x110](upload://tY562Weifz9EWup3XmYRF2CthqJ.png) batch atmul attrs not set the attrs type. --- [Visit Topic](https://discuss.tvm.apache.org/t/op-testing-single-op-ir-testing-batch-matmul/12049/2) to respond. You are receiving this bec

[Apache TVM Discuss] [Questions] [op testing] single op ir testing batch matmul

2022-02-09 Thread chenugray via Apache TVM Discuss
df_parsed = tvm.parser.parse_expr( ''' fn (%p0527: Tensor[(16, 256, 256), float32], %p1361: Tensor[(16, 64, 256), float32]) -> Tensor[(16, 256, 64), float32] { nn.batch_matmul(%p0527, %p1361, transpose_b=True) /* ty=Tensor[(16, 256, 64), float32] */ } ''') the code above

[Apache TVM Discuss] [Questions] [Pytorch] The inference results of tvm and pytorch are inconsistent

2022-02-09 Thread wyc0926 via Apache TVM Discuss
I think in order to ensure the accuracy of the model, rounding is necessary. ``` diff --git a/include/tvm/topi/nn/pooling.h b/include/tvm/topi/nn/pooling.h index c81c7cda7..467d2f5d8 100644 --- a/include/tvm/topi/nn/pooling.h +++ b/include/tvm/topi/nn/pooling.h @@ -386,7 +386,7 @@ inline Tensor a