Holger Schurig wrote:
That would be libsh, which Nick mentioned already :) I think that there still probably needs to be some research done to see how much this might help; i.e. is there going to be a lot of latency or overhead in moving data to/from the GPU, and making things happen? One characteristic of what we're trying to do here is that, for translators, we're performing lots of operations (encode/decode) on very small data sets; So, in most cases, the encode operation might be: 1) copy 20ms of PCM audio to GPU-accessible memory (unless it can access main memory?) 2) Call the GPU routines. 3) The GPU does the encoding. 4) Wait for the GPU to complete. 5) copy back the encoded 20ms data. As opposed to the Primary CPU equivalent: C1) The CPU does the encoding. This needs to be done 50 times/sec/call... So, while (3) might be much faster than C1, 1+2+3+4+5 might not.. Of course, there may be ways to alleviate this, like coding many calls together or something, to exploit the GPU parallelism, etc, but there's still a real possibility that this might not be a win at all.. Certainly an interesting research project, though. -SteveK |
_______________________________________________ Asterisk-Dev mailing list [email protected] http://lists.digium.com/mailman/listinfo/asterisk-dev To UNSUBSCRIBE or update options visit: http://lists.digium.com/mailman/listinfo/asterisk-dev
