I agree that in general the quality of the parallelism in most codes is rather low, unfortunately. But it is hard to proof that much can be gained when the quality would be improved.

Let me elaborate. When developing an app. that needs to run fast, one first needs to look at using the best algorithm to get the job done. While implementing the algorithm, attention must be paid to the app being stable (no use in having a fast app which crashes the whole time). And finally you start optimizing. But while using a better algorithm might give you a 50% boost, performance increases due to code-optimization are generally only marginal. Basically, changes early in the development process will have a big effect on performance while changes late in the dev.process. will have minor effects.

For instance, I wonder if any real-life application got a 50% boost by just changing the switch (and the corresponding MPI implementation). Or, what is exactly the speedup observed by switching from switch A to switch B on a real-life application?

toon


Patrick Geoffray wrote:
Hi Mark,

Mark Hahn wrote:
all these points are accurate to some degree, but give a sad impression
of the typical MPI programmer. how many MPI programmers are professionals,
rather than profs or grad students just trying to finish a calculation?
I don't know, since I only see the academic side.

I think that the sample of MPI codes or traces that I have seen so far is a good representation of the academic, labs and commercial sides. It's pretty bad. I am sure they are many reasons, but a few come to mind:

* a lot of codes in academia and at the labs are written directly by the scientist, physicist, chemist, whatever. They are expert in their domain, but they don't know how to write good code. Doesn't matter if it's parallel or sequential, they don't know how to do it right. For their defense, they never really learned, and they are doing the best they can. However, they really should work with professional programmers. It's paradoxical that physicists would use the service of a statistician to help them make sense of their experimental data, but they don't want help for computer science. It's interesting to note that there has always been this push from high in the food chain to bypass the human computer science expertise: it was automagic compilers (OpenMP, HPF and family) in the past, it's "high-productivity" languages now.

* In the commercial side, the codes are quite old, at least in their design. You can see traces of port from SHMEM to MPI, with Barriers a-lot-and-often. You see collective communications done by hand, I guess because the implementation of the collectives sucked at the time. You see an shameful amount of unexpected messages, the kind where the receive is just a little too late, typical from a code that was designed for a slow network, relatively. In short, it looks like they minimize the investment in code maintenance.


for academics, time-to-publish is the main criterion, which doesn't necessarily mean well-designed or tuned code. taking a significant

I don't know if time is really the constraint here. For grads students, sure, but I would not think that more time would help with profs. A good programing book maybe, but they are too proud to read those :-)

Patrick


--
Toon Knapen

------------------------------------------------
Check out our training program on acoustics
and register on-line at http://www.fft.be/?id=35
_______________________________________________
Beowulf mailing list, [email protected]
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to