Could you provide these parameters?
b = tvm.var('b') # batch size -> 12 or n n = tvm.var('n') # sequence length -> 512 h = tvm.var('h') # number of heads -> ? m = tvm.var('m') # hidden dimension -> 768 w = tvm.var('w') # window size -> 512 w_upper = tvm.var('w_upper') # window size to the right of the word. Should be `0` or `w` -> ? padding = tvm.var('padding') # padding -> ? transpose_t1 = tvm.var('transpose_t1') # t1 should be transposed -> True / False t1d3 = tvm.var('t1d3') # last dimension of t1 -> ? t3d3 = tvm.var('t3d3') # last dimension of t3 (the result tensor) -> ? --- [Visit Topic](https://discuss.tvm.ai/t/developing-a-faster-schedule-for-longformers-kernel/6367/5) to respond. You are receiving this because you enabled mailing list mode. To unsubscribe from these emails, [click here](https://discuss.tvm.ai/email/unsubscribe/68c35d21cc36e3ea2fe966a01e3e6d9d41e9f1b00a838c29d3c4b9c3c03c9052).