Re: [Beowulf] 9 traits of the veteran Unix admin

2011-02-18 Thread Bill Rankin
> Besides, I only HAVE around 20 cases of bottles to fill, and if > I get more the administrative penalty of using our laundry room wall as > a beer storage unit will become, um, "severe". > > In other words, before I reach that point I will have to acknowledge > that I've passed a critical scaling

Re: [Beowulf] IBM's Watson on Jeopardy tonight

2011-02-16 Thread Bill Rankin
> ariel sabiguero yawelak wrote: > > A dead company used to say "the > computer is the network", and for our brains it seems so. Actually for HPC and esp. clusters that (IMNSHO) is even more true today. My desktop just serves as my window into that network. Loss of a CPU or an entire server I

RE: [Beowulf] RE: Storage - the end of RAID?

2010-11-01 Thread Bill Rankin
> Um, it's not really RAID 1 when the drives are in different servers. > Although there's not much point in arguing about that. > > -- greg My knee-jerk reaction to Greg's statement was going to be something snotty and along the lines of "what part of 'R' don't you understand?" ;-) But upon fu

RE: [Beowulf] China Wrests Supercomputer Title From U.S.

2010-10-29 Thread Bill Rankin
> Define "real" applications, Something that produces tangible, scientifically useful results that would not have otherwise been realized without the availability and capability of that machine. > but to give my guess at your question "But they didn't. Why?" > > One word - cost Well, that's

RE: [Beowulf] China Wrests Supercomputer Title From U.S.

2010-10-29 Thread Bill Rankin
Douglas: > > [...] > > What this machine does do is validate to some extent the continued > use and development of GPUs in an HPC/cluster setting. > > [...] > > Nvidia claims Tianhe-1A's 4.04 megawatts of CUDA GPUs and Xeon CPUs is > three times more power efficient than CPUs alone. The Nvidia p

RE: [Beowulf] Re: Interesting

2010-10-29 Thread Bill Rankin
> "Robert G. Brown" wrote: > > > I've lost stories I've > > written on paper, and a really cool poem that I wrote with a pen popular > > in the 70's that turned out to have ink that faded to clear over 20 > > year, with or without the help of ambient UV. I have spiral notebooks > > from graduate

RE: [Beowulf] China Wrests Supercomputer Title From U.S.

2010-10-29 Thread Bill Rankin
> [...]and if its users are ordinary or not. > > -- greg In my experience very few people in this business would ever be called "ordinary". :-) -b ___ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscripti

RE: [Beowulf] China Wrests Supercomputer Title From U.S.

2010-10-28 Thread Bill Rankin
I was just going to post the same thing, but with the HPCWire link instead. http://www.hpcwire.com/blogs/New-China-GPGPU-Super-Outruns-Jaguar-105987389.html A few comments: A 42% increase (2.5TF v. 1.75TF for jaguar) is not a "wide margin". If I did my math right, that represents about 7 month

RE: [Beowulf] Looking for references for parallelization and optimization

2010-10-28 Thread Bill Rankin
If you are looking for more theoretical approaches, there is always John Reif’s book: John Reif (ed), “Synthesis of Parallel Algorithms”, published by Morgan Kaufmann, Spring, 1993. http://portal.acm.org/citation.cfm?coll=GUIDE&dl=GUIDE&id=562546 It is a weighty tome

RE: [Beowulf] Interesting

2010-10-28 Thread Bill Rankin
> Go to the moon. Dig a really big hole > at one of the poles (to avoid thermal extremes) and build a bunker out > of fused glass three meters thick and one kilometer underground. Ahh, I see that RGB has now revealed the plans for his villainous moon-base lair from which he will launch his nefar

RE: [Beowulf] how Google warps your brain

2010-10-26 Thread Bill Rankin
Heading completely off-topic now, but the area of digital media and long-term archival/retrieval is something that I find very interesting. I'll leave it to Rob to somehow eventually tie this back into a discussion of COTs technology and HPC. > > It's interesting: I just got an iPad a few wee

RE: [Beowulf] how Google warps your brain

2010-10-21 Thread Bill Rankin
Good points by Jim, and while I generally try and avoid "me too" posts, I just wanted to add my two cents. In my previous life I worked on building a central HPC cluster facility at Duke. The single biggest impediment to creating this resource was actually trying to justify its expense and pu

Re: [Beowulf] A sea of wimpy cores

2010-09-17 Thread Bill Rankin
On Sep 17, 2010, at 7:39 AM, Hearns, John wrote: http://www.theregister.co.uk/2010/09/17/hotzle_on_brawny_and_wimpy_cores Interesting article (more of a letter really) - to be honest when I first scanned it I was not sure of what Holzle's actual argument was. To me, he omitted a lot of the de

RE: [Beowulf] Cluster Metrics? (Upper management view)

2010-08-23 Thread Bill Rankin
Michael Di Domenico wrote: > I think measuring a clusters success based on the number of jobs run > or cpu's used is a bad measure of true success. I would be more > inclined to consider a cluster a success by speaking with the people > who use it and find out not only whether they can use it eff

RE: [Beowulf] dollars-per-teraflop : any lists like the Top500?

2010-07-02 Thread Bill Rankin
> > Another reason why some vendors are willing to > > sell stuff at reduced prices to universities is > > for visibility. The thinking is that when grad > > students (finally) graduate and go off into > > industry, they'll want to buy the same stuff > > they used when they were students. I'm not

RE: [Beowulf] dollars-per-teraflop : any lists like the Top500?

2010-07-02 Thread Bill Rankin
David: > > This thread brings to mind the punch line from the second of the SNL > > "Citiwide Change Bank" ads. Here is a link to the transcript: > > > > http://snltranscripts.jt.org/88/88achangebank2.phtml Thank you, now my keyboard is covered in coffee. ;-) > H. > > I'd just returned

RE: [Beowulf] dollars-per-teraflop : any lists like the Top500?

2010-06-30 Thread Bill Rankin
> I think the money part will be difficult to get (it is like a > politically > incorrect question). Joe addressed this pretty well. For the large systems, it's almost always under NDA. > Nevertheless, you can split the money in two parts: purchase (which I > am sure > you will never get) and e

RE: [Beowulf] MPI Persistent Comm Question

2010-06-28 Thread Bill Rankin
Uhmmm, what is "MPI_Free()"? It does not appear to be part of the MPI2 standard. MPI_Free_mem() is part of the standard, but is used in conjunction with MPI_Alloc_mem() and doesn't seem to refer to what you are describing here. Is it a local procedure and if so, what does it do? -bill Fro

RE: [Beowulf] cluster scheduler for dynamic tree-structured jobs?

2010-05-17 Thread Bill Rankin
Andrew Piskorski wrote: > Folks, I could use some advice on which cluster job scheduler (batch > queuing system) would be most appropriate for my particular needs. > I've looked through docs for SGE, Slurm, etc., but without first-hand > experience with each one it's not at all clear to me whic

Re: [Beowulf] copying data between clusters

2010-03-05 Thread Bill Rankin
methods other people use... > > Just a general survey... > ___ > Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing > To change your subscription (digest mode or unsubscribe) visit > http://www.beowulf.org/mailman/list

Re: [Beowulf] scheduler recommendations for a HPC cluster

2009-10-06 Thread Bill Rankin
Hi Rahul, If you are looking at the commercial offerings, you may want to also consider PBS Professional from Altair Engineering. http://www.pbspro.com/ In the name of full disclosure, I work for the PBS division of Altair. Thanks, -bill On 10/06/2009 03:22 PM, Rahul Nabar wrote: Any str

Re: [Beowulf] Please help to setup Beowulf

2009-02-25 Thread Bill Rankin
On Thu, 2009-02-19 at 13:28 -0800, Bill Broadley wrote: > While true, I ran openpbs and pbs pro (via a free to .edu program) and they > were both terrible. [...] > Granted this was 5+ years ago, PBS Pro has undergone major changes and revisions over the last five years. Back in 03-04 it was muc

Re: [Beowulf] Please help to setup Beowulf

2009-02-19 Thread Bill Rankin
> The fact that I > never met a serious PBS shop that had not made local custom changes to > the source code also soured me on deploying it when I was putting such > things into conservative IT shops who were still new and fearful of > Linux. One thing to note on Chris's review is that the

Re: [Beowulf] 10th Annual Beowulf Bash: Announcement and sponsorship opportunity

2008-10-31 Thread Bill Rankin
d sponsorship > opportunity > > > Tenth Annual Beowulf Bash > And > LECCIBG > > November 17 2008 9pm at Pete's Dueling Piano Bar -- Bill Rankin, Ph.D. Altair Engineering, PBS Gridworks [EMAIL PROTECTED] ___ Be

Re: [Beowulf] Simple MPI programs hang

2008-03-20 Thread Bill Rankin
On Mar 19, 2008, at 7:31 PM, Gregg Germain wrote: Hi everyone, I've created a 2 node cluster running FC8. I've installed MPICH2 1.0.6pl on both (not NFS'd). The Master, Ragnar, is a 64 bit; olaf is a 32 bit. I set up the ring, and mpdtrace shows: $ mpdtrace -l Ragnar_37601 (192.168.0.2)

Re: [Beowulf] PVM on wireless...

2008-02-07 Thread Bill Rankin
On Feb 7, 2008, at 1:41 PM, Robert G. Brown wrote: On Thu, 7 Feb 2008, Bill Rankin wrote: I think that I managed to replicate your problem, Rob. Laptop running CentoOS5, pvm 3.4.5-7(rpm), wireless ethernet. Server running FC6, pvm 3.4.5-7(rpm) Ssh working fine in both directions, PVM_ROOT

Re: [Beowulf] PVM on wireless...

2008-02-07 Thread Bill Rankin
I think that I managed to replicate your problem, Rob. Laptop running CentoOS5, pvm 3.4.5-7(rpm), wireless ethernet. Server running FC6, pvm 3.4.5-7(rpm) Ssh working fine in both directions, PVM_ROOT and PVM_RSH set accordingly. Running "pvm" from the shell on the server and doing an "add

Re: [Beowulf] Re: PVM on wireless...

2008-02-07 Thread Bill Rankin
x27; prompt. Sorry I can't offer anything more. -bill On Feb 6, 2008, at 6:18 PM, Bill Rankin wrote: I have a home setup similar to yours - a WAP acting as a firewall, dhcp from a linux server. I have a spare laptop running CentOS, so I'll give it a check tonight to see if mine runs.

Re: [Beowulf] Re: PVM on wireless...

2008-02-06 Thread Bill Rankin
I have a home setup similar to yours - a WAP acting as a firewall, dhcp from a linux server. I have a spare laptop running CentOS, so I'll give it a check tonight to see if mine runs. Q1: to you have DHCP giving a static address to your laptop based upon it's MAC? Q2: have you tried this

Re: [Beowulf] PVM on wireless...

2008-02-06 Thread Bill Rankin
Hey Rob, Could it be a node naming issue where the wireless IP does not resolve to the same address as that used in the machinefile? I seem to recall a similar issue back when we PVM on machines with multiple network connections. Just a thought, -bill On Feb 6, 2008, at 10:40 AM, Robe

Re: [Beowulf] Re: Cheap SDR IB

2008-02-05 Thread Bill Rankin
There's definitely potential for better interconnects and game clusters (deja vu, we must have discussed this some 5-8 years ago). Yeah, and my experiences with 2ndL are highly negatory as a consequence. It is a bad cluster design. It does not scale. I have not puttered around in SL for

Re: [Beowulf] TIPC in a Beowulf?

2008-02-04 Thread Bill Rankin
Hey Mark, I'm looking forward to OpenMX - it's a message-passing layer amenable to ethernet, but well-suited for MPI. any OpenMX people care to comment? Do you have any links to the current status of this effort? All my Googling leads to links on a package (also called OpenMX) for nano-

Re: [Beowulf] For only NAMD users

2008-01-05 Thread Bill Rankin
On the NAMD website: http://www.ks.uiuc.edu/Research/namd/ If you look through the release notes: http://www.ks.uiuc.edu/Research/namd/2.6/notes.html towards the bottom of the document they have a note on running NAMD with some simple input files. Hope this helps, -bill On Jan 5, 2008,

Re: VM and performance (was Re: [Beowulf] best Linux distribution)

2007-10-09 Thread Bill Rankin
On Oct 9, 2007, at 9:51 AM, Joe Landman wrote: Jeffrey B. Layton wrote: The recent emails from rgb and Doug lead me to a question. Has anyone tested codes running under a VM versus running them "natively" on the hardware (native isn't a good word and I hope everyone gets my meaning)? The last

Re: [Beowulf] best linux distribution

2007-10-08 Thread Bill Rankin
Yes, we use it with good effect on our 500+ node cluster at Duke. It's currently running Centos-4. I think that the only issue is that some of our developers require newer releases of a couple packages, but it's easy enough to maintain a local yum repository with those packages. It's be

Re: [Beowulf] 32 nodes cluster price

2007-10-08 Thread Bill Rankin
On Oct 8, 2007, at 3:38 AM, Geoff Galitz wrote: I would argue that the situation you describe is a result of that particular RAID adapter or that particular make and model is just inappropriate (no offense) None taken. I should have been clearer on the point I was trying to make. First the

Re: [Beowulf] [AMD64] Gentoo or Fedora

2007-10-08 Thread Bill Rankin
On Oct 7, 2007, at 6:42 PM, Greg Lindahl wrote: On Sun, Oct 07, 2007 at 03:10:30PM -0700, Greg Lindahl wrote: Hm, elm doesn't compile anymore, I wonder if anyone will notice if I just delete it? Of course, my CEO noticed about 10 minutes later! I told him to use a real mailer, like mutt. ;

Re: [Beowulf] 32 nodes cluster price

2007-10-07 Thread Bill Rankin
On Oct 5, 2007, at 4:17 PM, Leif Nixon wrote: "Geoff Galitz" <[EMAIL PROTECTED]> writes: Why do you automatically distrust hardware raid? To some extent I share Mark's sentiment. I certainly trust the Linux kernel more than the firmware in a cheap raid controller. Let me offer up a somewh

Re: [Beowulf] MPI performance gain with jumbo frames

2007-06-13 Thread Bill Rankin
Doug and Jeff have good points (and some good links). On thing to also pay attention to is the CPU utilization during the bandwidth and application testing. We found that on our cluster (various Dells with built in GigE NICs) while we did not see huge differences in effective bandwidth, t

Re: [Beowulf] MPI application benchmarks

2007-05-07 Thread Bill Rankin
Toon Knapen wrote: Mark Hahn wrote: sure. the suggestion is only useful if the cluster is dedicated to a single purpose or two. for anything else, I really think that microbenchmarks are the only way to go. I'm not sure that I agree with this - there are just so many different micro be

Re: [Beowulf] Survey on beowulf.org -- and a drawing for a video iPod

2007-04-13 Thread Bill Rankin
I'm glad I'm not the only one that had that problem. For any relatively large site, you're going to end up with a mostly custom configuration, even if you start with a vendor's products. Ya, well, I just said exactly that in my own response. Question one is a text box, at least, not a

[Beowulf] OT: Looking for HPC Headhunter/Recruiter

2007-04-05 Thread Bill Rankin
cheduled RGB rant. -bill -- Bill Rankin, Ph.D. Director, Cluster and Grid Technology Group Center for Computational Science, Engineering and Medicine Duke University [EMAIL PROTECTED] ph: (919) 660-6561 fx: (919) 660-7070 ___ Beowulf mailing list,

Re: [Beowulf] LECCIBG Announcement (SC06)

2006-11-07 Thread Bill Rankin
On Nov 7, 2006, at 10:27 AM, Douglas Eadline wrote: This is heart of the cigar world. Walt Ligon and I have been training six years for this. Well Doug, for us relative cigar novices, what advice would you offer as to preparation for this event? I assume that there will be opportunity

Re: [Beowulf] Explanation of error message in MPICH-1.2.7

2006-10-16 Thread Bill Rankin
I often see this error when a MPI_Barrier() call is not placed in front of the MPI_Finalize(). One of the processes exits early and MPICH doesn't like that too much. -b On Oct 6, 2006, at 3:50 PM, Jeffrey B. Layton wrote: Afternoon cluster fans, I'm working with a CFD code using the PG

Re: [Beowulf] [OT] HPC and University IT - forum/mailing list?

2006-08-16 Thread Bill Rankin
I was unclear - communication is great, and "outreach" is one of the things I spend a fair amount of time on. that's not what I meant by big-A, though - I meant people who sit behind a big desk behind a big secretary who decide that today is the day to solve all our problems by standardi

Re: [Beowulf] [OT] HPC and University IT - forum/mailing list?

2006-08-16 Thread Bill Rankin
On Aug 15, 2006, at 8:50 PM, Joe Landman wrote: [very good insight snipped] In order to get more support for little a, and I would argue for little/big S (end user support/machine support), you can't just be a cost center. You need to be a profit center. This forces a particular range of

Re: [Beowulf] [OT] HPC and University IT - forum/mailing list?

2006-08-15 Thread Bill Rankin
interesting number to know. Thanks for the feedback, -bill -- Bill Rankin, Ph.D. Director, Cluster and Grid Technology Group Center for Computational Science, Engineering and Medicine Duke University [EMAIL PROTECTED] ph: (919) 660-6561 fx: (919) 660-7070 _

[Beowulf] [OT] HPC and University IT - forum/mailing list?

2006-08-14 Thread Bill Rankin
of clusters and other HPC resources within a university environment? Would there be any interest in participating in such a forum? Thanks in advance, -bill -- Bill Rankin, Ph.D. Director, Cluster and Grid Technology Group Center for Computational Science, Engineering and Medicine Duke

[Beowulf] Multidimensional FFTs

2006-02-28 Thread Bill Rankin
uld be grateful. Some specifics that I am interested in would be a good comparison of different interconnects on overall performance, as this will have a significant impact on the design of their cluster. Thanks, -bill -- Bill Rankin, Ph.D. Director, Cluster and Grid Technology G

Re: [Beowulf] distributions

2006-02-02 Thread Bill Rankin
er compilers can all be Good Things. Commercial third party apps are probably the biggest sticking point. -b -- Bill Rankin, Ph.D. Director, Cluster and Grid Technology Group Center for Computational Science, Engineering and Medicine Duke University [EMAIL PROTECTED] ph: (919) 660-656

Re: [Beowulf] distributions

2006-02-02 Thread Bill Rankin
On Feb 2, 2006, at 8:04 AM, Robert G. Brown wrote: What we do is use centos for servers (LAN/department servers, that also serve the cluster nodes with e.g. home and project space). We use FC-even revision numbers for desktops, cluster nodes, etc. And for a slightly different view of othe