Dear Peter W.

Once I raised question similar to yours, having my requirement somwhat
very similar to yours. Please see my email in Beowulf Archive with
title:

"Slection from processor choices; Requesting Giudence"

posted during the later half of June this year.

At the end of discussion, I (opted) came up with the following design:

Each node having two AMD Opteron Dual Core processors and 4GB of main
memory on Tyan Thunder Board with GiGE as the interconnect.

regards,
AMJAD ALI.
BZU, Multan, Pakistan.



On 9/28/06, Peter Wainwright <[EMAIL PROTECTED]> wrote:
Please enlighten a baffled newbie:

Now there are motherboards with 8 sockets; quad-core processors; and
clusters
with as many nodes as you can shake a stick at.
It seems there are at least 3 dimensions for expansion.  What (in your
opinion) is the right tradeoff between more cores, more processors and
more
individual compute nodes?

In particular, I am thinking of in-house parallel finite difference /
finite element codes,
parallel BLAS, and maybe some commercial Monte-Carlo codes (the last
being an
embarrassingly parallel problem).

I have been set the task of building our first cluster for these
applications.
Our existing in-house codes run on an SGI machine with a parallelizing
compiler.
They would need to be ported to use MPI on a cluster.  However, I do not
understand
what happens when you have multi-processor/multi-core nodes in a
cluster.  Do you
just use MPI (with each thread using its own non-shared memory) or is
there any
way to do "mixed-mode" programming which takes advantage of shared
memory within a
node (like, an MPI/OpenMP hybrid?).

Peter Wainwright






_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf




_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to