That means this PR will generate the C code for CUTLASS?
---
[Visit
Topic](https://discuss.tvm.apache.org/t/if-target-c-how-to-execute-the-c-program/11519/8)
to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://
Yes, different autoTVM/autoscheduler runs on the same network can yield
different implementations.
If you know your model doesnt change, you would autoschedule once and save the
log files of the optimized implementation. If for some reason you need to
recompile your model you would tell TVM t
Ok, thank you for your reply.
---
[Visit
Topic](https://discuss.tvm.apache.org/t/if-target-c-how-to-execute-the-c-program/11519/7)
to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://discuss.tvm.apache.org/ema
Then you could consider the BYOC flow. You could refer to a recent effort that
integrates NVIDIA CUTLASS with BYOC flow and the C codegen:
https://github.com/apache/tvm/pull/9261
---
[Visit
Topic](https://discuss.tvm.apache.org/t/if-target-c-how-to-execute-the-c-program/11519/6)
to respo
Because I want to add a new backend in TVM, but there is no LLVM support for
our backend. So I plan to compile the model to c code(it only supports c).
---
[Visit
Topic](https://discuss.tvm.apache.org/t/if-target-c-how-to-execute-the-c-program/11519/5)
to respond.
You are receiving this
Why you need to get a model in a C program? In general it's not effective for
compilers to generate human readable C code unless this is the only format that
the corresponding vendor tool chain accepts. For example, TVM does have a
C-like codegen, but it generates OpenCL code for non-NVIDIA GP
Thank you for your reply. I'm confused whether tvm can compile the a tensorflow
model to a c program .
---
[Visit
Topic](https://discuss.tvm.apache.org/t/if-target-c-how-to-execute-the-c-program/11519/3)
to respond.
You are receiving this because you enabled mailing list mode.
To unsub
TVM deals with these in the Relay IR directly. For example, the IR with NCHW16c
and NCHW4c may look like:
```
%1 = nn.conv2d(...) // output layout: NCHW16c
%2 = layout_transform(%1, "NCHW4c") // output layout: NCHW4c
...
```
When compiling the above IR, `layout_tranform` is just an operator li
relay.build doesn't generate C code for execution. It generates LLVM IR (in the
case of CPU) and compiles directly to the executable binary.
---
[Visit
Topic](https://discuss.tvm.apache.org/t/if-target-c-how-to-execute-the-c-program/11519/2)
to respond.
You are receiving this because you
I have been trying to study how TVM does layout transformation during runtime
(eg. NHWC16c -> NHWC4c, etc.). Where in the source code is the required data
copy or move of the data tensor handled?
Also, where is the same for the weights tensor handled?
Is it in the `CopyDataFromTo` function of
Are the times reported by `tvm.runtime.profiler_vm.VirtualMachineProfiler`'s
`profile()` function the actual times observed on running inference on the
target device?
---
[Visit
Topic](https://discuss.tvm.apache.org/t/time-reported-by-the-virtual-machine-profiler/11522/1)
to respond.
Y
11 matches
Mail list logo