+1, be faster will be better, can't wait for any seconds
--
Reply to this email directly or view it on GitHub:
https://github.com/apache/tvm/issues/16368#issuecomment-1899964801
You are receiving this because you are subscribed to this thread.
Message ID:
I have working on unity branch for a couple of months ^_^
--
Reply to this email directly or view it on GitHub:
https://github.com/apache/tvm/issues/16368#issuecomment-1899967390
You are receiving this because you are subscribed to this thread.
Message ID:
with the popular of LLMs,NLP models is becoming more and more big,although can
be quantized, its still hard to deploy one LLM model in one GPU Ram. running
LLMs on multi-Host-multi-GPUs may be the usage solution right now.
I'm wondering, if we can deploy large model on multiply GPUs with tvm-u
Thanks! this really helped me, perfect!
---
[Visit
Topic](https://discuss.tvm.apache.org/t/elegant-and-efficient-ways-needed-in-debugging-tvm-c-files/15480/3)
to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https:/