Re: [dmlc/tvm] [QNN] [RFC] - Adding QNN operators with simple lowering (#3617)

2019-08-01 Thread 黎明灰烬
> Thanks @jackwish and @FrozenGene I understand your points. > > This can be treated as optimization then. If the input zero point is zero OR > if the input and output quantization params are same, don't cast, directly > apply maxpool. Generally, we would like to keep QNN APIs generic. So, if >

Re: [dmlc/tvm] [QNN] [RFC] - Adding QNN operators with simple lowering (#3617)

2019-08-01 Thread Animesh Jain
Thanks @jackwish and @FrozenGene I understand your points. This can be treated as optimization then. If the input zero point is zero OR if the input and output quantization params are same, don't cast, directly apply maxpool. Generally, we would like to keep QNN APIs generic. So, if MxNet for s

Re: [dmlc/tvm] [QNN] [RFC] - Adding QNN operators with simple lowering (#3617)

2019-08-01 Thread 黎明灰烬
> Thanks @jackwish for confirming the python lowering looks good. > > For max pooling, we used casting, because we have to subtract the zero point > from the quantized tensor. That subtract needs to happen in higher precision > than (u)int8. Correct me if I am wrong. To me, for Pooling operator

Re: [dmlc/tvm] [QNN] [RFC] - Adding QNN operators with simple lowering (#3617)

2019-08-01 Thread Zhao Wu
For TFLite model's average_pool / max_pool, which doesn't have input_zero_point. So, for TFLite, we don't need to subtract zero_point. MXNet have? -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/

Re: [dmlc/tvm] [QNN] [RFC] - Adding QNN operators with simple lowering (#3617)

2019-08-01 Thread Animesh Jain
Thanks @jackwish for confirming the python lowering looks good. For max pooling, we used casting, because we have to subtract the zero point from the quantized tensor. That subtract needs to happen in higher precision than (u)int8. Correct me if I am wrong. -- You are receiving this because yo

Re: [dmlc/tvm] [QNN] [RFC] - Adding QNN operators with simple lowering (#3617)

2019-08-01 Thread 黎明灰烬
I guess the `cast` is not needed in max pooling? Anyway, simply logic for compile operators looks good. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/issues/3617#issuecomment-517517945

Re: [dmlc/tvm] [QNN] [RFC] - Adding QNN operators with simple lowering (#3617)

2019-08-01 Thread Animesh Jain
@FrozenGene Can you please review #3627 -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/issues/3617#issuecomment-517491952

Re: [dmlc/tvm] [RFC] Add AVX512VNNI support for TVM (#3388)

2019-08-01 Thread Tianqi Chen
@jianyuh please act on the review comments @were please https://docs.tvm.ai/contribute/code_review.html#approve-and-request-changes-explicitly -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/

Re: [dmlc/tvm] [RFC] [Contrib] [Runtime] Minimal runtime (~12kb .text on ARMv7/x86) for subset of TVM models (#3567)

2019-08-01 Thread Tianqi Chen
@ajtulloch please look into the CI error and see if we can fix it. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3567#issuecomment-517433044

Re: [dmlc/tvm] [RFC] Add AVX512VNNI support for TVM (#3388)

2019-08-01 Thread Zhao Wu
If we have time, we could investigate why we couldn't achieve 252GFlops even more. Only 73% hardware efficiency means we have much work could dive. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/

Re: [dmlc/tvm] [RFC] Add AVX512VNNI support for TVM (#3388)

2019-08-01 Thread Jianyu Huang
@FrozenGene @tqchen @anijain2305 @llyfacebook @were Ping for review. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3388#issuecomment-517164224

Re: [dmlc/tvm] [RFC] Add AVX512VNNI support for TVM (#3388)

2019-08-01 Thread Jianyu Huang
Similar to @anijain2305 's PR (https://github.com/dmlc/tvm/pull/3516), currently we disable the AVX512 VNNI test in this PR. Posted the question on tensorize failure in https://discuss.tvm.ai/t/workaround-for-tensorize-failure/3577. -- You are receiving this because you are subscribed to this