It's nice to see some Apple centric notes here. Excellent writeup.

I've been involved in building and supporting Apple clusters running Sun Grid Engine for years now, dating back to the time when the first Xserves were released and we got some notoriety for using Apple 2nd generation iPods as bootable firewire drives for auto-imaging our cluster nodes.

Our apple cluster experience is more focused on batch-style compute farming rather than "true HPC" but I can toss some comments into the mix here - I've commented down below on some of the points that were raised in the write-up.

It's sad to hear from people attending SuperComputing that Apple did not have a booth. The consensus reported back to me was that "Apple has nothing to show in the HPC space ..." and that sort of goes along with what we've been seeing with Apple turning back from enterprise sales and focusing almost exclusively on the consumer market. Sad to hear -- we've been waiting on refreshed Xserves for way too long now. At this point I'd guess that the product may be dead or discontinued.


On Nov 20, 2007, at 2:19 PM, Kragen Javier Sitaker wrote:

This is not strictly about Beowulfs, but it is probably of interest to
their users.

My friend Joe's team from Indiana University just fielded a MacOS
cluster for the Supercomputing '07 Cluster Challenge.  His experiences
weren't that great; I encouraged him to jot something quick down so that
other people could benefit from his hard-won lessons.

There's more information about the challenge at
http://sc07.supercomp.org/?pg=clusterchlng.html&pp=conference.html.

----- Forwarded message from Kragen Javier Sitaker <[EMAIL PROTECTED]> -----

From: Kragen Javier Sitaker <[EMAIL PROTECTED]>
To: kragen-fw
Subject: Joe Blaylock's notes on running a MacOS cluster, Nov. 2007

 Disordered thoughts on using MacOS X for HPC.

By Joe Blaylock, 2007-11.


   Recollections:

   * we were the first people to ever try that particular combination:
     Tiger on Xeons with Intel's ICC 10 compiler suite and MKL linear
     algebra libraries. Blazing new territory is never easy.
   * We didn't use XGrid or Apple's cluster management stuff, only
     Server Admin and ARD.
   * Pov-Ray was easy; OpenMPI was easy; using Myrinet over 10Gig
     Ethernet was easy
   * GAMESS was more challenging, but we got it working somewhat. We
     still don't know how to run jobs of type ccsd(t), which require
     System V shared memory.
   * We never got POP to work.
* Apparently, ICC 10 has some bugs. There were several times when we
     were trying to use it to build, IIRC, GAMESS or POP, and it would
     give illegal instruction errors during compile. Or it would build
     a binary that we would run, and then it would do something
     horrible (like hang the machine (probably a bug interaction
     between icc and MacOSX).
   * OpenDirectory doesn't seem ready for prime time. It's pretty easy
     to set up, but it's unreliable and mysterious. In MacOS X, there
     seems to be a fundamental disconnect between things in the CLI
     world and things in the GUI world. Setting something up in one
     place won't necessarily be reflected in the other place. I'm sure
     that this is all trivial, if you're a serious Darwin user. But
     none of us were. So for example, you set up your NFS exports in
     the Server Admin tool, rather than by editing /etc/exports. The
     Admin tool won't put anything into /etc/exports. So if you're on
the command line, how do you check what you're exporting? With the
     complexity of LDAP, this becomes a real problem. You set up
     accounts on your head node, and say to export that information.
     But perhaps you create an account, but can't log into it on a
     node. If you're ssh'd in from the outside, where do you check to
     see (from the command-line) what the authentication system is
     doing? Our local Mac guru couldn't tell us. And then you'd create
another account, and the first one would start working again. WTF?
   * This may be the most frustrating thing about working with OS X
     Server. The CLI is the redheaded stepchild, and lots of HPC is
     mucking about on the command-line. You can use VNC to connect to
     ARD (but only if a user is logged in on the desktop and running
     ARD!), but it's slow, and only provides desktop control, not
cluster management. ARD can then be run on the desktop, to provide
     desktop control of the nodes in the cluster, and some cluster
     management: run unix command everywhere, shut nodes down, etc.
     There were a handful of tasks which seemed important, but which I
couldn't figure out how to do on the command-line at all. The most
     heinous of these is adding and removing users to/from LDAP.

Prior to Apple releasing the Xserve, there were many things that required a GUI to accomplish. Right around the time they released the Xserves though the OS improved to the point where you could just about everything via the command line and/or a serial console.

ServerAdmin even has a CLI variant that can do everything that the GUI. I think 'serveradmin' along with 'networksetup' were the two main CLI tools we used over and over again.

OpenDirectory is more of a pain - there is a CLI tool but you also end up working with standard openldap commands and binaries.

The best reference we found (in addition to lots of blood, sweat and tears) was a PDF that Apple publishes called "Mac OS X Server Command Line Administration" which can be found here: http://www.apple.com/server/pdfs/Command_Line.pdf

All things considered we've found that working on headless Apple boxes using a serial console is certainly possible, it's a bit slower than Linux due to the need to look up cryptic CLI commands in the PDF but it works for about 99% of the things we ever really needed to do. I think one GUI only thing that bit us once was the fact that during initial OS install if you want to software RAID your disks you can only do this via a GUI prompt.


   * Most of the time, I found it more convenient to use a 'for' loop
     that would ssh to nodes to run some command for me.

We use passwordless SSH and 'dsh' or 'pdsh' utilities on just about every Apple system we work on. There are many times when the "do something on N nodes" automated process is necessary. heh


* MacOS X lacks a way to do cpu frequency scaling. This killed us in
     the competition. We couldn't scale cores to save on our power
     budget, we could only leave them idle.
   * Being a Linux dude, I found having to have license keys for my
     operating systems, and (separately) my administration and
     management tools, to be odious in the extreme. Having to
     separately license ICC and IFORT and MKL just added frustration
     and annoyance.

This is not really a "secret" but Apple does not publish it easily and it took some internal contacts to discover ...

Apple is capable of producing what is called a "Watermarked Serial Number" that will license all servers in a cluster or a subnet. You still have to enter a serial number in the OS but having the single Watermark serial number allows you to use the same values which makes it scriptable or something you can bake into a netboot'ed image.

Watermarked serial numbers can not be requested by mere mortals. You have to request this from your sales rep and apparently there is some internal system that the sales rep can use to generate the watermarked serial number. This is the first thing we tell our Apple using friends and colleagues as it is a significantly nice thing to have.


   * We didn't make detailed performance comparisons between stuff
     built with the intel suite and things built with, e.g., the GNU
suite and GotoBLAS. We were too busy just trying to get everything
     to work. I'm sure that Intel produces better code under normal
     circumstances, but we had lots of cases where version 10 couldn't
     even produce viable binaries. So, make of that what you will.


   What I would recommend (if you were going to use MacOS X):

   * Learn Darwin, in detail. Figure out the CLI way to do everything,
     and do it. In fact, forget Mac OS X; just use Darwin. Learn the
     system's error codes, figure out how to manipulate fat binaries
(and how to strip them to make skinny ones), be able to manipulate
     users, debug the executing binaries, etc. Consider looking into
     the Apple disk imaging widget so you can boot the nodes diskless.


   What I would do differently (whether I stick with MacOS X or not):

   * diskless clients
   * Flash drive for head node
   * no GPUs
   * Get Serial Console set up and available, even if you don't use it
     routinely
   * CPU Frequency Scaling!!
   * many more, smaller cores. we had 36 at 3GHz. this was crazy. We
     were way power hungry.
   * Go to Intel 45nm dies.



Regards,
Chris



_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to