Re: [Beowulf] Opteron cooling specifications?

2006-05-01 Thread Bruce Allen
What breakers are not tested for is the trip time when operated at under its current rating. It is not uncommon especially as the breakers age or have been tripped repeatedly, to trip at less than 100% of its rating. I've replaced breakers many times that would trip when under 80-95% of its rate

Re: [Beowulf] 512 nodes Myrinet cluster Challanges

2006-05-01 Thread Robert G. Brown
On Mon, 1 May 2006, David Kewley wrote: For clusters with more than perhaps 16 nodes, or EVEN 32 if you're feeling masochistic and inclined to heartache: with all respect to rgb, I don't think size is a primary factor in cluster building/maintaining/etc effort. certainly it does eventually

[Beowulf] Re: 512 nodes Myrinet cluster Challanges (Walid)

2006-05-01 Thread MarkC Westwood
Walid Once you've surmounted the challenges of power and cooling and maintenance, you can start tackling the challenges of making your codes run efficiently on such a large cluster. And that's where the fun begins, and that's what makes us do it. Good luck Mark Westwood Principal Software Engi

[Beowulf] RARP request:MAC address match but no netmask match

2006-05-01 Thread stalin natesan
  hi all! i am using redhat 9 for my master node with via-rhine network card(in motherboard itself). When i am trying to boot slave node using floppy, everything goes fine, until RARP: Sending RARP requests. after sometimes the system reporting the PCI devices list and restarts. when i opened

Re: [Beowulf] Opteron cooling specifications?

2006-05-01 Thread Bari Ari
Mark Hahn wrote: sure, there's no downside to 220 afaikt. one thing I don't fully understand is where the 80% figure comes from. just a fudge-factor, and what if you wind up always using that extra capacity? presumably an L6-30 circuit is actually safe to run at 30... Safe to run sure, but i

Re: [Beowulf] 512 nodes Myrinet cluster Challanges

2006-05-01 Thread Vincent Diepeveen
Just measure the random ring latency of that 1024 nodes Myri system and compare. There is several tables around with the random ring latency. http://icl.cs.utk.edu/hpcc/hpcc_results.cgi To quote an official who sells huge Myri clusters, and i'm 100% sure he wants to keep anonymous: "you get wh

Re: [Beowulf] Opteron cooling specifications?

2006-05-01 Thread Mark Hahn
> > > 50u racks are also available (I've seen them in colo facilities)... > > > personally I don't like racking things that far over my head. > > > > I don't really care about that, myself - racking any serious number > > is a two-person job anyway. > > Two or more people for throughput -- in para

Re: [Beowulf] Opteron cooling specifications?

2006-05-01 Thread David Kewley
On Monday 01 May 2006 08:41, Mark Hahn wrote: > > 50u racks are also available (I've seen them in colo facilities)... > > personally I don't like racking things that far over my head. > > I don't really care about that, myself - racking any serious number > is a two-person job anyway. Two or more

RE: Electrical codes and plugs/sockets was Re: [Beowulf] Opteroncooling specifications?

2006-05-01 Thread Jim Lux
At 12:16 PM 5/1/2006, Michael Will wrote: And this is generally the message - short term it will work but long term continuous load should stay at 80%. There are different timeouts for breakers to be triggered at different over-80% levels. The overcurrent protection device should be able to ca

Re: [Beowulf] 512 nodes Myrinet cluster Challanges

2006-05-01 Thread David Kewley
> On Fri, 28 Apr 2006, David Kewley wrote: > > By the way, the idea of rolling-your-own hardware on a large cluster, > > and planning on having a small technical team, makes me shiver in > > horror. If you go that route, you better have *lots* of experience > > in clusters. and make very good deci

Re: [Beowulf] 512 nodes Myrinet cluster Challanges

2006-05-01 Thread David Kewley
On Monday 01 May 2006 13:37, Vincent Diepeveen wrote: > With so many nodes i'd go for either infiniband or quadrics, assuming the > largest partition also gets 512 nodes. > > Scales way better at so many nodes, as your software will need really a > lot of communications as you'll probably need quit

Re: [Beowulf] 512 nodes Myrinet cluster Challanges

2006-05-01 Thread Vincent Diepeveen
With so many nodes i'd go for either infiniband or quadrics, assuming the largest partition also gets 512 nodes. Scales way better at so many nodes, as your software will need really a lot of communications as you'll probably need quite a lot of RAM for the applications at all nodes. Of cour

RE: Electrical codes and plugs/sockets was Re: [Beowulf] Opteroncooling specifications?

2006-05-01 Thread Michael Will
And this is generally the message - short term it will work but long term continuous load should stay at 80%. There are different timeouts for breakers to be triggered at different over-80% levels. Michael -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of B

Re: Electrical codes and plugs/sockets was Re: [Beowulf] Opteron cooling specifications?

2006-05-01 Thread Bruce Allen
In our experience it's a very good idea to observe the "don't operate a circuit at more than 80% of rated capacity". We just had to replace a 150A 3-phase breaker which died after being operated a bit too close (within 10A) of its rated capacity for a number of months. Cheers, Bruce

Electrical codes and plugs/sockets was Re: [Beowulf] Opteron cooling specifications?

2006-05-01 Thread Jim Lux
At 08:41 AM 5/1/2006, Mark Hahn wrote: > > We've being switching cabinets in one of our datacenters to 220volt > service to support the sort of density we're seeing without running new > conductors. sure, there's no downside to 220 afaikt. one thing I don't fully understand is where the 80% fi

Re: [Beowulf] Opteron cooling specifications?

2006-05-01 Thread Mark Hahn
> 50u racks are also available (I've seen them in colo facilities)... > personally I don't like racking things that far over my head. I don't really care about that, myself - racking any serious number is a two-person job anyway. but taller racks make some sense to me, since open space above the