Closed #2692.
--
You are receiving this because you are subscribed to this thread.
Reply to this email directly or view it on GitHub:
https://github.com/dmlc/tvm/issues/2692#event-2231023869
Thanks @tqchen, glad to be a reviewer and looking forward to contributing more
in the future.
--
You are receiving this because you are subscribed to this thread.
Reply to this email directly or view it on GitHub:
https://github.com/dmlc/tvm/pull/3021#issuecomment-482881130
@MarisaKirisame, https://github.com/dmlc/tvm/pull/3448 was the leak @ajtulloch
was referring to.
--
You are receiving this because you are subscribed to this thread.
Reply to this email directly or view it on GitHub:
https://github.com/dmlc/tvm/issues/3423#issuecomment-506540526
@ziheng and @icemelon9, thanks very much. Looking forward to contributing more
in the future!
--
You are receiving this because you are subscribed to this thread.
Reply to this email directly or view it on GitHub:
https://github.com/dmlc/tvm/pull/3789#issuecomment-522184711
+1
--
You are receiving this because you commented.
Reply to this email directly or view it on GitHub:
https://github.com/dmlc/tvm/issues/4162#issuecomment-544779150
We have found a simple workaround in the case of concatenating 2D tensors
(currently our most common use case). By unrolling the last axis, llvm is smart
enough to generate vectorized code and the performance is even better than c
code in caffe2. For benchmark numbers, see
https://gist.github
Sounds good. Will do.
---
[Visit
Topic](https://discuss.tvm.ai/t/explore-optimizations-for-concat/2435/11) to
respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://discuss.tvm.ai/email/unsubscribe/488cbb585026277025