What relay expands to is memory copy. I want to avoid that. I want to have a
copy-less representation in TIR.
This should really be a no-op, but ends up copying everything.
```
import tensorflow as tf
import tvm
import tvm.relay
g = tf.Graph()
with g.as_default():
u = tf.unstack(tf.placeholder(dtype = 'int32', shape = (100,100)))
s = tf.stack(u)
m, p =
tvm.relay.frontend.tensorflow.from_tensorflow(g.as_graph_def(add_shapes=True))
g, m, p = tvm.relay.build_module.build(m, target='llvm')
m.save('pack.ll')
```
---
[Visit Topic](https://discuss.tvm.ai/t/tensor-arrays-in-tir/7135/4) to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://discuss.tvm.ai/email/unsubscribe/cb179268bb0b5b981ce694be3d323fd469d1c336583bbe196c2f4e6d9424e5e4).