I have mostly worked on pre-quantized models. So, I cant comment on the 
performance of Relay quantized model through ARM. There might be few missing 
pieces there.

I am planning to write a tutorial by next week on how to read pre-quantized 
models from TFLite. You can also try @masahi tutorial on PyTorch pre-quantized 
model import if you are blocked by this - 
https://github.com/apache/incubator-tvm/pull/5321

For eliminating "Cannot find config", you will have to tune the model on the 
device - 
https://tvm.apache.org/docs/tutorials/autotvm/tune_relay_arm.html#sphx-glr-tutorials-autotvm-tune-relay-arm-py





---
[Visit 
Topic](https://discuss.tvm.ai/t/is-there-any-speed-comparison-of-quantization-on-cpu/6256/43)
 to respond.

You are receiving this because you enabled mailing list mode.

To unsubscribe from these emails, [click 
here](https://discuss.tvm.ai/email/unsubscribe/a3f3092d139f5dbdc0e319828a2c170a21951ead7d2a92452a386e7d94432699).

Reply via email to