Re: [Beowulf] DC Power Dist. Yields 20%

2006-08-14 Thread Jim Lux
At 08:41 AM 8/11/2006, Geoff Jacobs wrote: Mark Hahn wrote: >> http://www.eweek.com/article2/0,1759,2000867,00.asp >> >> 20% is a lot, both in terms of consumption and cooling capacity. I'm > > the article says "up to 15%" at the facility level - did you get 20% > by figuring a savings at the rac

Re: [Beowulf] DC Power Dist. Yields 20%

2006-08-14 Thread Jim Lux
At 09:38 PM 8/10/2006, Mark Hahn wrote: http://www.eweek.com/article2/0,1759,2000867,00.asp 20% is a lot, both in terms of consumption and cooling capacity. I'm the article says "up to 15%" at the facility level - did you get 20% by figuring a savings at the rack level, too? afaikt, the artic

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Keith D. Underwood
Um, I haven't looked closely at Woodcrest lately, but everyone does remember that on a write, you have to fetch the cache line that you are writing to, right? So, if you have a 10 GB/s memory system, the most a copy should be able to do is: read the source at 3.3 GB/s read the destination at 3.3

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Joe Landman
Heh... Stuart Midgley wrote: > actually, latency determines bandwidth more than memory speed... see my > rant(?) from a little over a year ago > > http://www.beowulf.org/archive/2005-July/013294.html Someday the english language will be a complete impediment to understanding - Matt Austern We

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Vincent Diepeveen
Hi, Unlikely that a single core is the problem. Woodcrest 3Ghz we know little technical details though. A few scenario's for memory reads to L1 as i didn't see public technical data from intel yet. Most likely it has 1 port @ 2 cycles, as its 'father' the Banias (hmm bit bitter that near this

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread sNAAPS eLYK
Pardon my commenting on a discussion that's a little over my head, but if I understand the architechtural differences between Opteron and Pentium/Xeon, wouldn't the latency difference between them be attributed at least in part to the on-die memory controller? On 8/14/06, Stuart Midgley <[EMAIL P

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread sNAAPS eLYK
Pardon my commenting on a discussion that's a little over my head, but if I understand the architechtural differences between Opteron and Pentium/Xeon, wouldn't the latency difference between them be attributed at least in part to the on-die memory controller? On 8/14/06, Stuart Midgley <[EMAIL P

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Stuart Midgley
actually, latency determines bandwidth more than memory speed... see my rant(?) from a little over a year ago http://www.beowulf.org/archive/2005-July/013294.html prefetch etc. help, but the limiting factor is still latency. Hence the opterons have significantly higher real memory bandwidt

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Joe Landman
Hi Stu: Stu Midgley wrote: > sorry, forgot to reply all... don't you hate gmail's interface sometimes? > > > What is the memory latency of the woodcrest machines? Since memory > latency really determines your memory bandwidth. Hmmm... not for large block sequential accesses. You can prefetch

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Stu Midgley
sorry, forgot to reply all... don't you hate gmail's interface sometimes? What is the memory latency of the woodcrest machines? Since memory latency really determines your memory bandwidth. If Intel hasn't made any improvements in latency then the limited number of out-standing loads in the x8

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Richard Walsh
Joe Landman wrote: >4-threads > >Copy:6645.4125 0.0965 0.0963 0.0976 >Scale: 6994.6233 0.0916 0.0915 0.0917 >Add: 6373.0207 0.1508 0.1506 0.1509 >Triad: 6710.7522 0.1432 0.1431 0.1433 > >I may have

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Joe Landman
Jason Holmes wrote: Joe Landman wrote: I have it on good authority that with the other chipset (we have a Blackford here), we should see higher numbers. Not exceeding the Opteron 275 though. We have both a greencreek based (the one with the snoop filter.. I guess they're calling it the 50

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Jason Holmes
Joe Landman wrote: I have it on good authority that with the other chipset (we have a Blackford here), we should see higher numbers. Not exceeding the Opteron 275 though. We have both a greencreek based (the one with the snoop filter.. I guess they're calling it the 5000x now) and a blackf

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Vincent Diepeveen
Yes very dissappointing bandwidth. Probably the simple logics is that in a dual machine, 2 memory controllers is always faster than 1. http://www.cs.virginia.edu/stream/stream_mail/2005/0020.html 2x 2.8Ghz opteron - Function Rate (MB

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Jason Holmes
Mark Hahn wrote: numbers. It seems that the fortran version of the program with pathscale 2.3 (the numbers I reported above) reports different numbers than the C version of the program with pathscale 2.3: my numbers were from a random stream binary I had around, so I don't really even remembe

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Joe Landman
Mark Hahn wrote: kinda sucks, doesn't it? here's what I get for a not-new dual-275 with 8x1G PC3200 (I think): Function Rate (MB/s) RMS time Min time Max time Copy:5714.6837 0.0840 0.0840 0.0841 Scale: 5821.0766 0.0825 0.0825 0.

[Beowulf] EuroPVM/MPI 2006: Submissions for Late&Breaking Results

2006-08-14 Thread Joachim Worringen
Dear *, please allow me to post a final CALL FOR PAPERS for the EuroPVM/MPI conference (www.pvmmpi06.org): A new type of session introduced with EuroPVM/MPI 2006 are the "Late and Breaking Results": 2-page abstracts on the latest results and approaches can be submitted until September, 1st,

Re: [Beowulf] [OT] HPC and University IT - forum/mailing list?

2006-08-14 Thread Mark Hahn
Does anyone know of any mailing lists or forums that discuss the issues surrounding design/use/operation of clusters and other HPC resources within a university environment? Would there be any interest in participating in such a forum? I may be wrong, but I think a lot of the people on this l

Re: [Beowulf] [OT] HPC and University IT - forum/mailing list?

2006-08-14 Thread Joe Landman
Hi Bill Bill Rankin wrote: Hey gang, Sorry for the slightly OT post, but I figure that clusters have been such a staple of the higher ed HPC diet for so many years that this group is probably the best to ask. Does anyone know of any mailing lists or forums that discuss the issues surroundi

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Mark Hahn
numbers. It seems that the fortran version of the program with pathscale 2.3 (the numbers I reported above) reports different numbers than the C version of the program with pathscale 2.3: my numbers were from a random stream binary I had around, so I don't really even remember which compiler f

[Beowulf] [OT] HPC and University IT - forum/mailing list?

2006-08-14 Thread Bill Rankin
Hey gang, Sorry for the slightly OT post, but I figure that clusters have been such a staple of the higher ed HPC diet for so many years that this group is probably the best to ask. Does anyone know of any mailing lists or forums that discuss the issues surrounding design/use/operation of

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Jason Holmes
Mark Hahn wrote: Here's a 4 core Woodcrest system with 4 sticks of 2 GB 667 MHz FB-dimms: Function Rate (MB/s) Avg time Min time Max time Copy: 4813.2590 0.2663 0.2659 0.2675 Scale: 4826.2181 0.2656 0.2652 0.2662 Add:5244.0135 0.3664

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Mark Hahn
Here's a 4 core Woodcrest system with 4 sticks of 2 GB 667 MHz FB-dimms: Function Rate (MB/s) Avg time Min time Max time Copy: 4813.2590 0.2663 0.2659 0.2675 Scale: 4826.2181 0.2656 0.2652 0.2662 Add:5244.0135 0.3664 0.3661 0

Re: [Beowulf] Woodcrest Memory bandwidth

2006-08-14 Thread Jason Holmes
Here's a 4 core Woodcrest system with 4 sticks of 2 GB 667 MHz FB-dimms: Function Rate (MB/s) Avg time Min time Max time Copy: 4813.2590 0.2663 0.2659 0.2675 Scale: 4826.2181 0.2656 0.2652 0.2662 Add:5244.0135 0.3664 0.3661 0

[Beowulf] EuroPVM/MPI 2006: Early Registration Deadline ahead!

2006-08-14 Thread Joachim Worringen
*** *** *** EuroPVM/MPI 2006 *** *** www.pvmmpi06.org *** ***

Re: [Beowulf] gamers: evil or just useless? ;)

2006-08-14 Thread Douglas Eadline
> On Tue, Aug 08, 2006 at 06:41:14PM -0400, Mark Hahn wrote: > >> first GPUs, now NICs? http://www.bigfootnetworks.com/ appears to be >> an actual linux-based coprocessor marketed as an offload NIC for gamers. >> weird, but true! from their whitepaper, I can't tell whether their >> performance nu