I revise the input name in
[imagenet_test.py](https://github.com/Edgecortix-Inc/pytorch_quantization/blob/master/tvm_qnn_evaluation/imagenet_test.py)
as following:

But get the following error while execute for resnet18 model:

to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](h
New to TVM, still getting used to the best ways to express algorithms in
`te.compute`:
Is there a better way to express a 2x3 matrix inverse than writing a long
cascading `te.compute(...te.if_then_else((i,j) == (0,0),
te.if_then_else(...` or using hybrid script?
---
[Visit Topic](ht
Hopefully someone can correct me if I'm wrong, but I believe the tophub logs
are downloaded anytime you run relay's build(...).
So I believe the answer to you question, is just don't run the autotvm and just
go right to building.
with relay.build_config(opt_level=4):
graph, lib, params =
Thanks a lot. I've been playing around with this on a BERT model, but I'm
hitting some issues when calling `relay.build` with opt level 3. The target is
`cuda`. The error message looks like this:
```
unresolved intrinsic sqrt with return type float16x4
```
It comes from `codegen_c.cc`. Does t
Hi. I am trying to use the graph tuner module on CPU.
Normally, you would run autoTVM, pick the best configuration, then instantiate
the graph tuner:
executor = Tuner(graph, input_dict, records, target_op, target)
In many cases, [the default schedules already provided good
performance](
Yes,thanks again for your reply.
I just verified [
tutorial_eager.py](https://github.com/Edgecortix-Inc/pytorch_quantization/blob/master/tutorial_eager.py)
@torch-nightly(v1.6) @macbook pro, and get the 2-4x speed-up as the
[static_quantization_tutorial](https://pytorch.org/tutorials/advanced
Yes,thanks again for your reply.
---
[Visit
Topic](https://discuss.tvm.ai/t/is-there-any-speed-comparison-of-quantization-on-cpu/6256/9)
to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://discuss.tvm.ai/email/
I ended up not using `scan`. I used `extern` tensor instead, and I wrote the
generating function myself.
---
[Visit Topic](https://discuss.tvm.ai/t/scan-on-a-non-zero-axis/5996/3) to
respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails,
1. I don't have experience using QAT in Torch. I think post training
quantization is easier to work with. In any case, post training quantization
should be the first thing you should try. If you need extra accuracy, QAT may
help.
2. Yes. See
https://docs.tvm.ai/tutorials/frontend/deploy_qua
As the GPU vendor recommend "Tell the Compiler the Work-Group Size" by
```
__attribute__((reqd_work_group_size(X,Y,Z))).
```
in OpenCL kenrel function.
I can't get the work group size in codegen, Could you share how to get the work
group size?
---
[Visit
Topic](https://discuss.tvm.ai
Hello!
I wrote an op composed of four CUDA kernels, and now I want to optimize the op,
so I need to know the time ratio of the four kernels.
I tried nvprof but was unable to use it due to permission issues.
Is there a similar test function in TVM?
My current test code is as follows:
mod
15 matches
Mail list logo