Well Mark, don't give up!
I am not sure which one is your application domain, but if you require 24x7 computation, then you should not be hosting that at home. On the other hand, if you are not doing real computation and you just have a testbed at home, maybe for debugging your parallel applications or something similar, you might be interested in a virtualized solution. Several years ago, I used to "debug" some neural networks at home, but training sessions (up to two weeks of training) happened at the university.
I would suggest to do something like that.
You can always scale-down your problem in several phases and save the complete data-set / problem for THE RUN.

You are not being a heretic there, but suffering energy costs ;-)
In more places that you may believe, useful computing nodes are being replaced just because of energy costs. Even in some application domains you can even loose computational power if you move from 4 nodes into a single quad-core (i.e. memory bandwidth problems). I know it is very nice to be able to do everything at home.. but maybe before dropping your studies or working overtime to pay the electricity bill, you might want to reconsider the fact of collapsing your phisical deploy into a single virtualized cluster. (or just dispatch several threads/processes in a single system). If you collapse into a single system you have only 1 mainboard, one HDD, one power source, one processor (physically speaking), .... and you can achieve almost the performance of 4 systems in one, consuming the power of.... well maybe even less than a single one. I don't want to go into discussions about performance gain/loose due to the variation of the hardware architecture. Invest some bucks (if you haven't done that yet) in a good power source. Efficiency of OEM unbranded power sources is realy pathetic. may be 45-50% efficiency, while a good power source might be 75-80% efficient. Use the energy for computing, not for heating your house. What I mean is that you could consider just collapsing a complete "small" cluster into single system. If your application is CPU-bound and not I/O bound, VMware Server could be an option, as it is free software (unfortunately not open, even tough some patches can be done on the drivers). I think it is not possible to publish benchmarking data about VMware, but I can tell you that in long timescales, the performance you get in the host OS is similar than the one of the guest OS. There are a lot of problems related to jitter, from crazy clocks to delays, but if your application is not sensitive to that, then you are Ok. Maybe this is not a solution, but you can provide more information regarding your problem before quitting...

my 2 cents....

ariel

Mark Kosmowski escribió:
At some point there a cost-benefit analysis needs to be performed.  If
my cluster at peak usage only uses 4 Gb RAM per CPU (I live in
single-core land still and do not yet differentiate between CPU and
core) and my nodes all have 16 Gb per CPU then I am wasting RAM
resources and would be better off buying new machines and physically
transferring the RAM to and from them or running more jobs each
distributed across fewer CPUs.  Or saving on my electricity bill and
powering down some nodes.

As heretical as this last sounds, I'm tempted to throw in the towel on
my PhD studies because I can no longer afford the power to run my
three node cluster at home.  Energy costs may end up being the straw
that breaks this camel's back.

Mark E. Kosmowski

From: "Jon Aquilina" <[EMAIL PROTECTED]>

not sure if this applies to all kinds of senarios that clusters are used in
but isnt the more ram you have the better?

On 6/30/08, Vincent Diepeveen <[EMAIL PROTECTED]> wrote:
Toon,

Can you drop a line on how important RAM is for weather forecasting in
latest type of calculations you're performing?

Thanks,
Vincent


On Jun 30, 2008, at 8:20 PM, Toon Moene wrote:

Jim Lux wrote:
Yep.  And for good reason.  Even a big DoD job is still tiny in Nvidia's
scale of operations. We face this all the time with NASA work.
 Semiconductor manufacturers have no real reason to produce special purpose
or customized versions of their products for space use, because they can
sell all they can make to the consumer market. More than once, I've had a
phone call along the lines of this:
"Jim: I'm interested in your new ABC321 part."
"Rep: Great. I'll just send the NDA over and we can talk about it."
"Jim: Great, you have my email and my fax # is..."
"Rep: By the way, what sort of volume are you going to be using?"
"Jim: Oh, 10-12.."
"Rep: thousand per week, excellent..."
"Jim: No, a dozen pieces, total, lifetime buy, or at best maybe every
year."
"Rep: Oh...<dial tone>"
{Well, to be fair, it's not that bad, they don't hang up on you..

Since about a year, it's been clear to me that weather forecasting (i.e.,
running a more or less sophisticated atmospheric model to provide weather
predictions) is going to be "mainstream" in the sense that every business
that needs such forecasts for its operations can simply run them in-house.

Case in point:  I bought a $1100 HP box (the obvious target group being
teenage downloaders) which performs the HIRLAM limited area model *on the
grid that we used until October 2006* in December last year.

It's about twice as slow as our then-operational 50-CPU Sun Fire 15K.

I wonder what effect this will have on CPU developments ...

--
Toon Moene - e-mail: [EMAIL PROTECTED] - phone: +31 346 214290
Saturnushof 14, 3738 XG  Maartensdijk, The Netherlands
At home: http://moene.indiv.nluug.nl/~toon/
Progress of GNU Fortran: http://gcc.gnu.org/ml/gcc/2008-01/msg00009.html

_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf


--
Jonathan Aquilina
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to