In message from "Tom Elken" <[EMAIL PROTECTED]> (Thu, 24 Apr 2008 09:31:16 -0700):
I have up to 1453 MB/s for 4MB message size ... on osu_bw test
w/Mellanox DDR IB
(Mellanox version of OFED-1.2, w/binary mvapich-0.9.9); OpenMPI
-1.2.2-1 gives even a
bit more (1470 MB/s - more exactly, 1469.753328, 1469.447179, 1469.977840 for 3 subsequent test runs).

The SC'07 message of D.K.Panda
http://mvapich.cse.ohio-state.edu/publications/sc07_mpich2_bof.pdf
inform us about 1405 MB/s.

Is this throughput difference the result of MPI-2 vs MPI implementation or should I beleive that this difference (about 4% for my mvapich vs mvapich2 at SC'07 ) is not significant - in the sense that it is simple because of some measurement errors (inaccuracies)?

The way to see if there is a real throughput difference between a MPI-2 implementation and a MPI-1 implementation is to measure it on your pair of machines.

:-) Of course - but I've the problem w/mvapich2 (from binary Mellanox/ofed-1.2) setting. When I try to run mpdboot (/etc/mpd.conf contains the same MPD_SECRETWORD оn both nodes; MV2_DEFAULT_DAPL_PROVIDER=ib0), I see

mpdboot -v -n 2 -f /where/is/mpihosts
mpdroot: perror msg: No such file or directory
running mpdallexit on <node1_shortname>
LAUNCHED mpd on  <node1_shortname> via
RUNNING: mpd on <node1_shortname>
LAUNCHED mpd on <node2_FQDN> via <node1_shortname> RUNNING: mpd on <node2_FQDN>
=================================================
/var/log/messages contains:

Apr 22 21:20:53 <node1_shortname> python2.4: mpdallexit: mpd_uncaught_except_tb handling: exceptions.
TypeError: not all arguments converted during string formatting
/usr/mpi/gcc/mvapich2-0.9.8-12/bin/mpdlib.py 899 __init__ mpd_print(1,'forked process failed; status=' % status) /usr/mpi/gcc/mvapich2-0.9.8-12/bin/mpdallexit.py 44 mpdallexit conSock = MPDConClientSock(mpdroot=mpdroot,secretword=parmdb['MPD_SECRETWORD']) /usr/mpi/gcc/mvapich2-0.9.8-12/bin/mpdallexit.py 59 ? mpdallexit()
Apr 22 21:20:53  <node1_shortname> mpd: mpd starting; no mpdid yet
Apr 22 21:20:53 <node1_shortname> mpd: mpd has mpdid=<node1_shortname>_40611 (port=40611) Apr 22 21:21:01 <node1_shortname> kernel: ib0: multicast join failed for ff12:601b:ffff:0000:0000:0001:ff22:e50d, status -22
Apr 22 21:21:33 c5ws7 last message repeated 2 times
                                       etc
=====================================================

... and I don't understand (even from strace output :-)) which
file want mpdboot/mpdroot :-(


Comparing your results to published results are difficult because nearly
all the variables need to be the same for the comparison to be valid.
Variables like which of the following were used in the two tests:
- Mellanox IB DDR adapter
- PCIe interface type
- CPU model and speed
- PCIe chipset
- OFED version, ...

Certainly the MPI flavor and version is important, but it is not, in
general, the most important of these factors.
Note for example these two results on the OSU MVAPICH web pages:

MVAPICH2  1-sided put throughput, measured with osu_bw:
1405 MB/s:  ConnectX DDR, PCIe x8, EM64T 2.33 GHz quad-core CPU
http://mvapich.cse.ohio-state.edu/performance/mvapich2/em64t/MVAPICH2-em
64t-gen2-ConnectX-DDR-1S.shtml

1481 MB/s:  MT25208 HCA silicon, PCIe x8, Intel Xeon 3.6 Ghz, EM64T

http://mvapich.cse.ohio-state.edu/performance/mvapich2/em64t/MVAPICH2-em
64t-gen2-DDR-1S.shtml

Both are DDR IB adapters.  ConnectX is the newer silicon.  But because
of system differences, the older adapter is faster, in this case.

Thanks for this reference ! I thought that on my more old HCA hardware (Infinihost III Lx PCI-e x8 MHGS18-XTC), more old CPU/mobo/... (Opteron 246/2 Ghz/...), more old Linux, ofed and mvapich/mvapich2 versions I must obtain more lower throughput value ...

Mikhail


-Tom



Mikhail Kuzminsky
Computer Assistance to Chemical Research Center
Zelinsky Institute of Organic Chemistry
Moscow
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf


_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to