Hi Joe,
Can that 9% difference be due to the Intel capability to overclock one core and turn the others off?
Or is does this Intel feature require manual switch somewhere?
Thank  you,
Kevin


Joe Landman wrote:
Hi folks:

Thought you might like to see this. I rewrote the interior loop for our Riemann Zeta Function (rzf) example for SSE2, and ran it on a Nehalem and on a Shanghai. This code is compute intensive. The inner loop which had been written as this (some small hand optimization, loop unrolling, etc):

    l[0]=(double)(inf-1 - 0);
    l[1]=(double)(inf-1 - 1);
    l[2]=(double)(inf-1 - 2);
    l[3]=(double)(inf-1 - 3);
    p_sum[0] = p_sum[1] = p_sum[2] = p_sum[3] = zero;
    for(k=start_index;k>end_index;k-=unroll)
       {
          d_pow[0] = l[0];
          d_pow[1] = l[1];
          d_pow[2] = l[2];
          d_pow[3] = l[3];

          for (m=n;m>1;m--)
           {
             d_pow[0] *=  l[0];
             d_pow[1] *=  l[1];
             d_pow[2] *=  l[2];
             d_pow[3] *=  l[3];
           }
          p_sum[0] += one/d_pow[0];
          p_sum[1] += one/d_pow[1];
          p_sum[2] += one/d_pow[2];
          p_sum[3] += one/d_pow[3];

          l[0]-=four;
          l[1]-=four;
          l[2]-=four;
          l[3]-=four;
      }
    sum = p_sum[0] + p_sum[1] + p_sum[2] + p_sum[3] ;

has been rewritten as

    __m128d __P_SUM = _mm_set_pd1(0.0);        // __P_SUM[0 ... VLEN] = 0
    __m128d __ONE = _mm_set_pd1(1.);   // __ONE[0 ... VLEN] = 1
    __m128d __DEC = _mm_set_pd1((double)VLEN);
    __m128d __L   = _mm_load_pd(l);

    for(k=start_index;k>end_index;k-=unroll)
       {
          __D_POW       = __L;

          for (m=n;m>1;m--)
           {
             __D_POW    = _mm_mul_pd(__D_POW, __L);
           }

__P_SUM = _mm_add_pd(__P_SUM, _mm_div_pd(__ONE, __D_POW));

          __L           = _mm_sub_pd(__L, __DEC);

      }

    _mm_store_pd(p_sum,__P_SUM);

    for(k=0;k<VLEN;k++)
     {
       sum += p_sum[k];
     }

The two codes were run on a Nehalem 3.2 GHz (desktop) processor, and a Shanghai 2.3 GHz desktop processor. Here are the results

    Code        CPU    Freq (GHz)    Wall clock (s)
    ------        -------    -------------    --------------

base Nehalem 3.2 20.5 optimized Nehalem 3.2 6.72 SSE-ized Nehalem 3.2 3.37

    base        Shanghai 2.3        30.3
optimized Shanghai 2.3 7.36 SSE-ized Shanghai 2.3 3.68 These are single thread, single core runs. Code scales very well (is one of our example codes for the HPC/programming/parallelization classes we do).

I found it interesting that they started out with the baseline code performance tracking the ratio of clock speeds ... The Nehalem has a 39% faster clock, and showed 48% faster performance, which is about 9% more than could be accounted for by clock speed alone. The SSE code performance appears to be about 9% different.

I am sure lots of interesting points can be made out of this (being only one test, and not the most typical test/use case either, such points may be of dubious value).

I am working on a Cuda version of the above as well, and will try to compare this to the threaded versions of the above. I am curious what we can achieve.

Joe


--
Kevin C. Abbey
System Administrator
Rutgers University - BioMaPS Institute

Email: kab...@biomaps.rutgers.edu


Hill Center - Room 259
110 Frelinghuysen Road
Piscataway, NJ  08854
Phone and Voice mail: 732-445-3288
Wright-Rieman Laboratories Room 201
610 Taylor Rd.
Piscataway, NJ  08854-8087
Phone: 732-445-2069
Fax: 732-445-5958

_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to