Hi Sayantan, We have installed OFED 1.2, and our two problems have gone! Now there is neither suspending processes and nor inconsistent communication times: PACKAGE SIZE 512 BYTES 1.76 PACKAGE SIZE 4096 BYTES 13.83
These were Our test: 512: 29.434 4096: 16.209 with OFED 1.1. Thanks and regards, Tahir Malas Bilkent University Electrical and Electronics Engineering Department Phone: +90 312 290 1385 > -----Original Message----- > From: Sayantan Sur [mailto:[EMAIL PROTECTED] > Sent: Tuesday, June 12, 2007 6:09 PM > To: Tahir Malas > Cc: [EMAIL PROTECTED]; beowulf@beowulf.org; > [EMAIL PROTECTED]; 'Ozgur Ergul' > Subject: Re: [mvapich-discuss] Two problems related to slowness and > TASK_UNINTERRUPTABLE process > > Hi Tahir, > > Thanks for sharing this data and your observations. It is interesting. > We have a more recent release, MVAPICH-0.9.9 which is available from our > website (mvapich.cse.ohio-state.edu) as well as with OFED-1.2 > distribution. Could you please try out our newer release and see if the > results change/remain the same? > > Thanks, > Sayantan. > > Tahir Malas wrote: > > Hi all, > > We have an 8 dual quad-core node HP cluster connected via Infiniband. We > use > > Voltaire DDR cards and 24-port switch. We also use OFED 1.1 and MVAPICH > > 0.9.7. We have two interesting problems that we could not overcome yet: > > > > 1. In our test program which mimics the communications in our code, the > > nodes are paired as follows: (0 and 1), (2 and 3), (4 and 5), (6 and 7). > We > > perform one to one communications between these pairs of nodes > > simultaneously. We use blocking MPI send and receive commands to > communicate > > an integer array of various sizes. In addition, we consider different > > numbers of processes: > > (a) 1 process per node, 8 processes overall: One link is established > between > > the pairs of nodes. > > (b) 2 process per node, 16 processes overall: Two links are established > > between the pairs of nodes. > > (c) 4 process per node, 32 processes overall: Four links are established > > between the pairs of nodes. > > (d) 8 process per node, 64 processes overall: Eight links are > established > > between the pairs of nodes. > > > > We obtain logical timings, except for the following interesting > comparison: > > > > For 32 processes (4 process per node), the arrays with 512-Byte size are > > communicated slower than the 4096-Byte size arrays. For both of them, we > > send/receive 1,000,000 arrays and take the average to find the time per > > package. Only package size changes. We have made many trials and > confirmed > > this abnormal case is persistent. More specifically, communication of > > 4k-Byte packages are 2 times faster than the communication of 512-Byte > > packages. > > > > The OSU bandwidth and latency test around these points shows: > > Byte MB/s > > 256 417.53 > > 512 592.34 > > 1024 691.02 > > 2048 857.35 > > 4096 906.04 > > 8192 1022.52 > > Time (usec) > > 256 4.79 > > 512 5.48 > > 1024 6.60 > > 2048 8.30 > > 4096 11.02 > > So this behavior does not seem reasonable to us. > > > > 2. SOMETIMES, after the test with overall 32 processes, one of the four > > processes at node3 hangs in TASK_UNINTERRUPTABLE "D" state. Hence, the > test > > program shows a "done." and waits for sometime. We can neither kill the > > process nor soft reboot the node. We have to wait for that process to > > terminate, which can last long. > > > > Does anybody have some comments in these issues? > > Thanks in advance, > > Tahir Malas > > Bilkent University > > Electrical and Electronics Engineering Department > > > > > > > > _______________________________________________ > > mvapich-discuss mailing list > > [EMAIL PROTECTED] > > http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss > > > > > -- > http://www.cse.ohio-state.edu/~surs > _______________________________________________ Beowulf mailing list, Beowulf@beowulf.org To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf