i am wondering if there is any chance to introduce a quick way to compatible 
with dynamic shapes? 
as @cloudhan mentioned, TensorRT can let user set necessary input dimensions at 
runtime, and auto compute other tensors' shape:
[Working With Dynamic 
Shapes](https://docs.nvidia.com/deeplearning/tensorrt/developer-guide/index.html#work_dynamic_shapes)

and i noticed, relay can parse models with dynamic shapes, but will failed at 
relay.build() or vm.compile()
so, can we have some feature like:
`
mod, params = relay.frontend.from_tensorflow(...)  
mod.get_tensor_by_name('input:0').set_shapes((...))  
mod.auto_compute_shapes()  
...  
relay.build(mod, target)  
`

to achieve the full support and optimization of dynamic shapes seems to be a 
huge project, and i found a lot users show their interesting and concern about 
this topic. i think maybe a little further step could be quickly done can be 
helpful.

@kevinthesun , i'm a newbie in tvm, maybe what i thought is too simple as a 
matter of course. just want to help, appreciate~

-- 
You are receiving this because you are subscribed to this thread.
Reply to this email directly or view it on GitHub:
https://github.com/apache/tvm/issues/4118#issuecomment-789559705

Reply via email to