@yangjunpro +1 on your work. As Relay is also planning to support dynamic shape(https://github.com/dmlc/tvm/issues/3042), we might not need to directly handle step 5(relay does the JIT/bucketing under the hood). We are also thinking that is it reasonable to do the opposite, the main runtime is tvm runtime, and fallback to TensorFlow for unsupported ops, would it yield a solution that has lower memory footprint? Maybe we can open a discussion thread to further discuss(on both)? Since we don't really have an action item yet. cc @yongwww @zhiics @icemelon9
-- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/issues/3059#issuecomment-485538857
