Kernels running on the GPU require all memory accesses to be within a thread or
a block. The file you are looking does not do any thread binding. I suggest
looking at this tutorial:
https://tvm.apache.org/docs/tutorials/optimize/opt_conv_cuda.html
---
[Visit
Topic](https://discuss.tvm.ap
Because the tutorial only tunes conv2d tasks:
tasks = autotvm.task.extract_from_program(
mod["main"], target=target, params=params,
ops=(relay.op.get("nn.conv2d"),)
)
Although I don't think missing one dense schedule will hurt the performance a
lot, you can modify the code to
@comaniac Hmm, not sure if you're not the right person for this question, but
I'd really appreciate it if you could provide some hints or pointers to others.
---
[Visit
Topic](https://discuss.tvm.apache.org/t/cannot-find-config-for-target-llvm-keys-cpu/8066/2)
to respond.
You are receivi
Thanks for the pointer!
---
[Visit
Topic](https://discuss.tvm.apache.org/t/graph-optimization-not-exist-in-gpu/8051/11)
to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://discuss.tvm.apache.org/email/unsubscri
```
with tvm.transform.PassContext(opt_level=3):
lib =relay.build_module.build(mod, target=target, params=params)
```
The Relay build process eventually calls the graph optimization pipeline.
---
[Visit
Topic](https://discuss.tvm.apache.org/t/graph-optimization-not-exist-in-gpu/8051/1