On Fri, 5 Dec 2008, Eugen Leitl wrote:


(Well, duh).

Good article, though, thanks.

Of course the same could have been written (and probably was) back when
dual processors came out sharing a single memory bus, and for every
generation since.  The memory lag has been around forever -- multicores
simply widen the gap out of step with Moore's Law (again).

Intel and/or AMD people on list -- any words you want to say about a
"road map" or other plan to deal with this?  In the context of ordinary
PCs the marginal benefit of additional cores after (say) four seems
minimal as most desktop users don't need all that much parallelism --
enough to manage multimedia decoding in parallel with the OS base
function in parallel with "user activity".  Higher numbers of cores seem
to be primarily of interest to H[A,PC] users -- stacks of VMs or server
daemons, large scale parallel numerical computation.  In both of these
general arenas increasing cores/processor/memory channel beyond a
critical limit that I think we're already at simply ensures that a
significant number of your cores will be idling as they wait for
memory access at any given time...

   rgb


http://www.spectrum.ieee.org/nov08/6912

Multicore Is Bad News For Supercomputers

By Samuel K. Moore

Image: Sandia

Trouble Ahead: More cores per chip will slow some programs [red] unless
there’s a big boost in memory bandwidth [yellow

With no other way to improve the performance of processors further, chip
makers have staked their future on putting more and more processor cores on
the same chip. Engineers at Sandia National Laboratories, in New Mexico, have
simulated future high-performance computers containing the 8-core, 16‑core,
and 32-core microprocessors that chip makers say are the future of the
industry. The results are distressing. Because of limited memory bandwidth
and memory-management schemes that are poorly suited to supercomputers, the
performance of these machines would level off or even decline with more
cores. The performance is especially bad for informatics
applications—data-intensive programs that are increasingly crucial to the
labs’ national security function.

High-performance computing has historically focused on solving differential
equations describing physical systems, such as Earth’s atmosphere or a
hydrogen bomb’s fission trigger. These systems lend themselves to being
divided up into grids, so the physical system can, to a degree, be mapped to
the physical location of processors or processor cores, thus minimizing
delays in moving data.

But an increasing number of important science and engineering problems—not to
mention national security problems—are of a different sort. These fall under
the general category of informatics and include calculating what happens to a
transportation network during a natural disaster and searching for patterns
that predict terrorist attacks or nuclear proliferation failures. These
operations often require sifting through enormous databases of information.

For informatics, more cores doesn’t mean better performance [see red line in
“Trouble Ahead”], according to Sandia’s simulation. “After about 8 cores,
there’s no improvement,” says James Peery, director of computation,
computers, information, and mathematics at Sandia. “At 16 cores, it looks
like 2.” Over the past year, the Sandia team has discussed the results widely
with chip makers, supercomputer designers, and users of high-performance
computers. Unless computer architects find a solution, Peery and others
expect that supercomputer programmers will either turn off the extra cores or
use them for something ancillary to the main problem.

At the heart of the trouble is the so-called memory wall—the growing
disparity between how fast a CPU can operate on data and how fast it can get
the data it needs. Although the number of cores per processor is increasing,
the number of connections from the chip to the rest of the computer is not.
So keeping all the cores fed with data is a problem. In informatics
applications, the problem is worse, explains Richard C. Murphy, a senior
member of the technical staff at Sandia, because there is no physical
relationship between what a processor may be working on and where the next
set of data it needs may reside. Instead of being in the cache of the core
next door, the data may be on a DRAM chip in a rack 20 meters away and need
to leave the chip, pass through one or more routers and optical fibers, and
find its way onto the processor.

In an effort to get things back on track, this year the U.S. Department of
Energy formed the Institute for Advanced Architectures and Algorithms.
Located at Sandia and at Oak Ridge National Laboratory, in Tennessee, the
institute’s work will be to figure out what high-performance computer
architectures will be needed five to 10 years from now and help steer the
industry in that direction.

“The key to solving this bottleneck is tighter, and maybe smarter,
integration of memory and processors,” says Peery. For its part, Sandia is
exploring the impact of stacking memory chips atop processors to improve
memory bandwidth.

The results, in simulation at least, are promising [see yellow line in
“Trouble Ahead

_______________________________________________
Beowulf mailing list, [email protected]
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf


Robert G. Brown                        http://www.phy.duke.edu/~rgb/
Duke University Dept. of Physics, Box 90305
Durham, N.C. 27708-0305
Phone: 1-919-660-2567  Fax: 919-660-2525     email:[EMAIL PROTECTED]

_______________________________________________
Beowulf mailing list, [email protected]
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to