On Dec 12, 2008, at 8:56 AM, Eugen Leitl wrote:
http://www.goodgearguide.com.au/article/270416/inside_tsubame_-
_nvidia_gpu_supercomputer?fp=&fpid=&pf=1
Inside Tsubame - the Nvidia GPU supercomputer
Tokyo Tech University's Tsubame supercomputer attained 29th ranking
in the
new Top 500, thanks in part to hundreds of Nvidia Tesla graphics
cards.
Martyn Williams (IDG News Service) 10/12/2008 12:20:00
When you enter the computer room on the second floor of Tokyo
Institute of
Technology's computer building, you're not immediately struck by
the size of
Japan's second-fastest supercomputer. You can't see the Tsubame
computer for
the industrial air conditioning units that are standing in your
way, but this
in itself is telling. With more than 30,000 processing cores
buzzing away,
the machine consumes a megawatt of power and needs to be kept cool.
1000000 watt / 77480 gflop = 12.9 watt per gflop.
If you run double precision codes on this box it is a big energy
waster IMHO.
(of course it's very well equipped for all kind of crypto codes using
that google library).
Vincent
Tsubame was ranked 29th-fastest supercomputer in the world in the
latest Top
500 ranking with a speed of 77.48T Flops (floating point operations
per
second) on the industry-standard Linpack benchmark.
While its position is relatively good, that's not what makes it so
special.
The interesting thing about Tsubame is that it doesn't rely on the raw
processing power of CPUs (central processing units) alone to get
its work
done. Tsubame includes hundreds of graphics processors of the same
type used
in consumer PCs, working alongside CPUs in a mixed environment that
some say
is a model for future supercomputers serving disciplines like material
chemistry.
Graphics processors (GPUs) are very good at quickly performing the
same
computation on large amounts of data, so they can make short work
of some
problems in areas such as molecular dynamics, physics simulations
and image
processing.
"I think in the vast majority of the interesting problems in the
future, the
problems that affect humanity where the impact comes from
nature ... requires
the ability to manipulate and compute on a very large data set," said
Jen-Hsun Huang, CEO of Nvidia, who spoke at the university this
week. Tsubame
uses 680 of Nvidia's Tesla graphics cards.
Just how much of a difference do the GPUs make? Takayuki Aoki, a
professor of
material chemistry at the university, said that simulations that
used to take
three months now take 10 hours on Tsubame.
Tsubame itself - once you move past the air-conditioners - is split
across
several rooms in two floors of the building and is largely made up of
rack-mounted Sun x4600 systems. There are 655 of these in all, each
of which
has 16 AMD Opteron CPU cores inside it, and Clearspeed CSX600
accelerator
boards.
The graphics chips are contained in 170 Nvidia Tesla S1070 rack-
mount units
that have been slotted in between the Sun systems. Each of the 1U
Nvidia
systems has four GPUs inside, each of which has 240 processing
cores for a
total of 960 cores per system.
The Tesla systems were added to Tsubame over the course of about a
week while
the computer was operating.
"People thought we were crazy," said Satoshi Matsuoka, director of
the Global
Scientific Information and Computing Center at the university.
"This is a ¥1
billion (US$11 million) supercomputer consuming a megawatt of
power, but we
proved technically that it was possible."
The result is what university staff call version 1.2 of the Tsubame
supercomputer.
"I think we should have been able to achieve 85 [T Flops], but we
ran out of
time so it was 77 [T Flops]," said Matsuoka of the benchmarks
performed on
the system. At 85T Flops it would have risen a couple of places in
the Top
500 and been ranked fastest in Japan.
There's always next time: A new Top 500 list is due out in June
2009, and
Tokyo Institute of Technology is also looking further ahead.
"This is not the end of Tsubame, it's just the beginning of GPU
acceleration
becoming mainstream," said Matsuoka. "We believe that in the world
there will
be supercomputers registering several petaflops in the years to
come, and we
would like to follow suit."
Tsubame 2.0, as he dubbed the next upgrade, should be here within
the next
two years and will boast a sustained performance of at least a
petaflop (a
petaflop is 1,000 teraflops), he said. The basic design for the
machine is
still not finalized but it will continue the heterogeneous
computing base of
mixing CPUs and GPUs, he said.
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf