[Apache TVM Discuss] [Questions] [BUG Report] auto dense large gpu schedule

2022-03-15 Thread chenugray via Apache TVM Discuss
Take this code for example: import numpy as np import tvm from tvm.autotvm.tuner import XGBTuner from tvm import relay, autotvm import pytest def test_dense_autotvm(): target = tvm.target.cuda() batch, in_dim, out_dim = 16384, 768, 768 data_shape =

[Apache TVM Discuss] [Questions] [Performance] can inference optimized away softmax

2022-02-24 Thread chenugray via Apache TVM Discuss
i saw the comment in nn.softmax This operator can be optimized away for inference for now, the bert performance bottleneck is related with softmax. what's the meaning of this comment,how to optimize away this op. the ir may like below: %1579 = fn (%p0218: Tensor[(128, 12, 128, 128), fl

[Apache TVM Discuss] [Questions] Any way to extract tir function from relay?

2022-02-10 Thread chenugray via Apache TVM Discuss
i wanna find way to automative single op accuracy testing. I use another way to extract single op ir definition, then run the single op ir func def. not using tir. --- [Visit Topic](https://discuss.tvm.apache.org/t/any-way-to-extract-tir-function-from-relay/11907/3) to respond. You are

[Apache TVM Discuss] [Questions] [op testing] single op ir testing batch matmul

2022-02-09 Thread chenugray via Apache TVM Discuss
[[BUG]fix batch matmul not set attrs_type_key, when using tvm.parse.parse_expr will raise error.](https://github.com/apache/tvm/pull/10209) --- [Visit Topic](https://discuss.tvm.apache.org/t/op-testing-single-op-ir-testing-batch-matmul/12049/4) to respond. You are receiving this because

[Apache TVM Discuss] [Questions] [op testing] single op ir testing batch matmul

2022-02-09 Thread chenugray via Apache TVM Discuss
ok, the op Registry do have some problem. ![image|690x110](upload://tY562Weifz9EWup3XmYRF2CthqJ.png) batch atmul attrs not set the attrs type. --- [Visit Topic](https://discuss.tvm.apache.org/t/op-testing-single-op-ir-testing-batch-matmul/12049/2) to respond. You are receiving this bec

[Apache TVM Discuss] [Questions] [op testing] single op ir testing batch matmul

2022-02-09 Thread chenugray via Apache TVM Discuss
df_parsed = tvm.parser.parse_expr( ''' fn (%p0527: Tensor[(16, 256, 256), float32], %p1361: Tensor[(16, 64, 256), float32]) -> Tensor[(16, 256, 64), float32] { nn.batch_matmul(%p0527, %p1361, transpose_b=True) /* ty=Tensor[(16, 256, 64), float32] */ } ''') the code above

[Apache TVM Discuss] [Questions] What if the result is not correct?

2022-01-09 Thread chenugray via Apache TVM Discuss
cause tvm leave no origin network layers infomation in the tvm graph, so how can i use dump data compared with orgin network layers? for exmaple, the bert-large has 2000+ ops, but which op related to origin layer is hard to figure out. when you face accuracy problem, you dump the data and com

[Apache TVM Discuss] [Questions] How to compare layers with origin network

2022-01-07 Thread chenugray via Apache TVM Discuss
cause tvm leave no origin network layers infomation in the tvm graph, so how can i use dump data compared with orgin network layers? for exmaple, the bert-large has 2000+ ops, but which op related to origin layer is hard to figure out. when you face accuracy problem, you dump the data and com

[Apache TVM Discuss] [Questions] What if the result is not correct?

2022-01-06 Thread chenugray via Apache TVM Discuss
> All the tensors will be saved as binary bytes in serialized format. The > result binary bytes can be loaded by the API “load_params”. how can we achive binary bytes in serialized format? does "./_tvmdbg_device_CPU_0/output_tensors.params" contains all layers outputs? then we get data, how to

[Apache TVM Discuss] [Questions] What if the result is not correct?

2022-01-06 Thread chenugray via Apache TVM Discuss
thx, could u give me some concrete example? --- [Visit Topic](https://discuss.tvm.apache.org/t/what-if-the-result-is-not-correct/11858/3) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click here](https://discuss.tvm.apache.o

[Apache TVM Discuss] [Questions] What if the result is not correct?

2022-01-06 Thread chenugray via Apache TVM Discuss
@tqchen Any Methods to debug to know which intermediate layer outputs is not correct. i seach the whole formu, but cant get answer. --- [Visit Topic](https://discuss.tvm.apache.org/t/what-if-the-result-is-not-correct/11858/1) to respond. You are receiving this because you enabled mailing

[Apache TVM Discuss] [Questions] How to extract tvm module

2021-12-30 Thread chenugray via Apache TVM Discuss
how to dump this graph? --- [Visit Topic](https://discuss.tvm.apache.org/t/how-to-extract-tvm-module/2167/22) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click here](https://discuss.tvm.apache.org/email/unsubscribe/bb6d1de

[Apache TVM Discuss] [Questions] Bert-large masked lm pre-quantization model build failed

2021-12-30 Thread chenugray via Apache TVM Discuss
from pytorch_pretrained_bert import BertForMaskedLM import torch def main(args): bert_model_origin = BertForMaskedLM.from_pretrained("bert-large-uncased") example_tensor = torch.randint(0, 100, (1, 256)) model_int8 = torch.quantization.quantize_dynamic(bert_m