What breakers are not tested for is the trip time when operated at under its
current rating. It is not uncommon especially as the breakers age or have
been tripped repeatedly, to trip at less than 100% of its rating. I've
replaced breakers many times that would trip when under 80-95% of its rate
On Mon, 1 May 2006, David Kewley wrote:
For clusters with more than perhaps 16 nodes, or EVEN 32 if you're
feeling masochistic and inclined to heartache:
with all respect to rgb, I don't think size is a primary factor in
cluster building/maintaining/etc effort. certainly it does eventually
Walid
Once you've surmounted the challenges of power and cooling and
maintenance, you can start tackling the challenges of making your
codes run efficiently on such a large cluster. And that's where the
fun begins, and that's what makes us do it.
Good luck
Mark Westwood
Principal Software Engi
hi all!
i am using redhat 9 for my master node with via-rhine network card(in motherboard itself). When i am trying to boot slave node using floppy, everything goes fine, until RARP: Sending RARP requests. after sometimes the system reporting the PCI devices list and restarts.
when i opened
Mark Hahn wrote:
sure, there's no downside to 220 afaikt. one thing I don't fully
understand is where the 80% figure comes from. just a fudge-factor,
and what if you wind up always using that extra capacity?
presumably an L6-30 circuit is actually safe to run at 30...
Safe to run sure, but i
Just measure the random ring latency of that 1024 nodes Myri system and
compare.
There is several tables around with the random ring latency.
http://icl.cs.utk.edu/hpcc/hpcc_results.cgi
To quote an official who sells huge Myri clusters, and i'm 100% sure he
wants to keep anonymous: "you get wh
> > > 50u racks are also available (I've seen them in colo facilities)...
> > > personally I don't like racking things that far over my head.
> >
> > I don't really care about that, myself - racking any serious number
> > is a two-person job anyway.
>
> Two or more people for throughput -- in para
On Monday 01 May 2006 08:41, Mark Hahn wrote:
> > 50u racks are also available (I've seen them in colo facilities)...
> > personally I don't like racking things that far over my head.
>
> I don't really care about that, myself - racking any serious number
> is a two-person job anyway.
Two or more
At 12:16 PM 5/1/2006, Michael Will wrote:
And this is generally the message - short term it will work
but long term continuous load should stay at 80%. There are different
timeouts for breakers to be triggered at different over-80% levels.
The overcurrent protection device should be able to ca
> On Fri, 28 Apr 2006, David Kewley wrote:
> > By the way, the idea of rolling-your-own hardware on a large cluster,
> > and planning on having a small technical team, makes me shiver in
> > horror. If you go that route, you better have *lots* of experience
> > in clusters. and make very good deci
On Monday 01 May 2006 13:37, Vincent Diepeveen wrote:
> With so many nodes i'd go for either infiniband or quadrics, assuming the
> largest partition also gets 512 nodes.
>
> Scales way better at so many nodes, as your software will need really a
> lot of communications as you'll probably need quit
With so many nodes i'd go for either infiniband or quadrics, assuming the
largest partition also gets 512 nodes.
Scales way better at so many nodes, as your software will need really a lot
of
communications as you'll probably need quite a lot of RAM for the
applications at all nodes.
Of cour
And this is generally the message - short term it will work
but long term continuous load should stay at 80%. There are different
timeouts for breakers to be triggered at different over-80% levels.
Michael
-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]
On Behalf Of B
In our experience it's a very good idea to observe the "don't operate a
circuit at more than 80% of rated capacity". We just had to replace a
150A 3-phase breaker which died after being operated a bit too close
(within 10A) of its rated capacity for a number of months.
Cheers,
Bruce
At 08:41 AM 5/1/2006, Mark Hahn wrote:
>
> We've being switching cabinets in one of our datacenters to 220volt
> service to support the sort of density we're seeing without running new
> conductors.
sure, there's no downside to 220 afaikt. one thing I don't fully
understand is where the 80% fi
> 50u racks are also available (I've seen them in colo facilities)...
> personally I don't like racking things that far over my head.
I don't really care about that, myself - racking any serious number
is a two-person job anyway. but taller racks make some sense to me,
since open space above the
16 matches
Mail list logo