Hi Masahi, I think dynamic split will be impossible, because the output tuple
will have a unknown size. The size of a tuple must be static in Relay.
However, it could be possible if the pytorch graph tells us the number of
outputs and we know it to be static.
---
[Visit
Topic](https://dis
The TensorRT execution we use in TVM is not asynchronous, so there is no need
to sync. `module.run()` won't return until inference is completed. Actually I
think run() is never asynchronous in TVM?
5ms is not an unreasonable inference time for mobilenet v2 with TensorRT on
xavier, although I