Hi @Hzfengsy, Thanks for your answer, that makes a lot of sense. 

However, what I am trying to communicate is that even with a new TIR and the 
ability to print it out at each step, the underlying scheduling language seems 
to be roughly the same in all of these examples (with the nice exception of 
removing the unnecessary `s[C]`). This is okay in all the examples in the 
thread that are very small. 

What I am suggesting is that this might be a perfect time to think about 
simplifying the scheduling language as well. For instance the example here: 
https://tvm.d2l.ai/chapter_gpu_schedules/matmul.html#blocked-matrix-multiplication-on-gpu
 is in some sense the simplest scheduling example of TVM, and it is still 
extremely complicated. It has roughly 20 local variables in scope all with one 
letter names and many overlapping mutations (without even using autotvm). 

I am curious about any ideas the new approach has to make this more robust, 
easier to extend, and understandable to non-experts?





---
[Visit 
Topic](https://discuss.tvm.apache.org/t/rfc-tensorir-a-schedulable-ir-for-tvm/7872/50)
 to respond.

You are receiving this because you enabled mailing list mode.

To unsubscribe from these emails, [click 
here](https://discuss.tvm.apache.org/email/unsubscribe/cfa418fa84383166a0b9981a44ba3c1d8cf6a19879e34b4802d8fe6fbf590a19).

Reply via email to