IMHO the hybris approach (MPI+threads) is interesting in case every
MPI-process has lots of local data.

yes. but does this happen a lot? the appealing case would be threads that make lots of heavy use of some large data, _but_
without needing synchronization/locking.  once you need locking
among the threads, message passing starts to catch up.

latter is simpler because it only requires MPI-parallelism but if the code
is memory-bound and every mpi-process has much of the same data, it will be
better to share this common data with all processes on the same cpu and thus
use threads intra-node.

what kind of applications behave like that? I agree that if your MPI app is keeping huge amounts of (static) data replicated in each rank,
you should rethink your design.

thanks, mark hahn.
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to