Please enlighten a baffled newbie: Now there are motherboards with 8 sockets; quad-core processors; and clusters with as many nodes as you can shake a stick at. It seems there are at least 3 dimensions for expansion. What (in your opinion) is the right tradeoff between more cores, more processors and more individual compute nodes?
In particular, I am thinking of in-house parallel finite difference / finite element codes, parallel BLAS, and maybe some commercial Monte-Carlo codes (the last being an embarrassingly parallel problem). I have been set the task of building our first cluster for these applications. Our existing in-house codes run on an SGI machine with a parallelizing compiler. They would need to be ported to use MPI on a cluster. However, I do not understand what happens when you have multi-processor/multi-core nodes in a cluster. Do you just use MPI (with each thread using its own non-shared memory) or is there any way to do "mixed-mode" programming which takes advantage of shared memory within a node (like, an MPI/OpenMP hybrid?). Peter Wainwright
signature.asc
Description: This is a digitally signed message part
_______________________________________________ Beowulf mailing list, Beowulf@beowulf.org To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf