> On Behalf Of Gilad Shainer
> 
> ... OSU has different benchmarks
> so you can measure message coalescing or real message rate. 

[ As a refresher for the wider audience , as Gilad defined earlier: " Message 
coalescing is when you incorporate multiple MPI messages in a single network 
packet."  And I agree with this definition :) ]

Gilad,  

Sorry for the delayed QLogic response on this.  I was on vacation when this 
thread started up.  But now that it has been revived, ... 

Which OSU benchmarks have message-coalescing built into the source?  

> Nowadays it seems that QLogic
> promotes the message rate as non coalescing data and I almost got
> bought
> by their marketing machine till I looked on at the data on the wire...
> interesting what the bits and bytes and symbols can tell you...

Message-coalescing has been done in benchmark source code, such as HPC 
Challenge's MPI RandomAccess benchmark.  In that case, coalescing is performed 
when the SANDIA_OPT2 define is turned on during the build.

More typically message coalescing is a feature of some MPIs and they use 
various heuristics for when it is active.

MVAPICH has an environment variable -- VIADEV_USE_COALESCE -- which can turn 
this feature on or off.  HP-MPI has coalescing heuristics on by default when 
using IB-Verbs, off by default when using QLogic's PSM.  Open MPI has enabled 
message-coalescing heuristics for more recent versions when running over IB 
verbs.

There is nothing wrong with message coalescing features in the MPI.  Only when 
you are trying to measure the raw message rate of the network adapter, it is 
best to not use message coalescing feature so you can measure what you set out 
to measure.  QLogic MPI does not have a message coalescing feature, and that is 
what we use to measure MPI message rate on our IB adapters.  We also measure 
using MVAPICH with it's message coalescing feature turned off, and get 
virtually identical message rate performance to that with QLogic MPI. 

I don't know what you were measuring on the wire, but with the osu_mbw_mr 
benchmark and QLogic MPI, for the small 1 to 8 byte message sizes where we 
achieve maximum message rate, each message is in its own 56 byte packet with
no coalescing.  I asked a couple of our engineers who have looked at a lot of 
PCIe traces to make sure of this.

Regards,
-Tom


> 
> 
> 
> -----Original Message-----
> From: beowulf-boun...@beowulf.org [mailto:beowulf-boun...@beowulf.org]
> On Behalf Of Greg Lindahl
> Sent: Friday, February 19, 2010 2:06 PM
> To: beowulf@beowulf.org
> Subject: Re: [Beowulf] Q: IB message rate & large core counts (per
> node)?
> 
> > Mellanox latest message rate numbers with ConnectX-2 more than
> > doubled versus the old cards, and are for real message rate -
> > separate messages on the wire. The competitor numbers are with using
> > message coalescing, so it is not real separate messages on the wire,
> > or not really message rate.
> 
> Gilad,
> 
> I think you forgot which side you're supposed to be supporting.
> 
> The only people I have ever seen publish message rate with coalesced
> messages are DK Panda (with Mellanox cards) and Mellanox.
> 
> QLogic always hated coalesced messages, and if you look back in the
> archive for this mailing list, you'll see me denouncing coalesced
> messages as meanless about 1 microsecond after the first result was
> published by Prof. Panda.
> 
> Looking around the Internet, I don't see any numbers ever published by
> PathScale/QLogic using coalesced messages.
> 
> At the end of the day, the only reason microbenchmarks are useful is
> when they help explain why one interconnect does better than another
> on real applications. No customer should ever choose which adapter to
> buy based on microbenchmarks.
> 
> -- greg
> (formerly employed by QLogic)
> 
> 
> _______________________________________________
> Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin
> Computing
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf
> 
> _______________________________________________
> Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin
> Computing
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf

_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to