Scott (and Michael and Carlos),

Thanks for your excellent feedback. That's the kind of enlightening feedback I was looking for. Interesting that the HBM on Fugaku exceeds the needs of the processor.

Prentice

On 6/16/21 2:23 PM, Scott Atchley wrote:

On Wed, Jun 16, 2021 at 1:15 PM Prentice Bisbal via Beowulf <beowulf@beowulf.org <mailto:beowulf@beowulf.org>> wrote:

    Did anyone else attend this webinar panel discussion with AMD
    hosted by
    HPCWire yesterday? It was titled "AMD HPC Solutions: Enabling Your
    Success in HPC"

    https://www.hpcwire.com/amd-hpc-solutions-enabling-your-success-in-hpc/
    <https://www.hpcwire.com/amd-hpc-solutions-enabling-your-success-in-hpc/>

    I attended it, and noticed there was no mention of AMD supporting
    AVX512, so during the question and answer portion of the program, I
    asked when AMD processors will support AVX512. The answer given,
    and I'm
    not making this up, is that AMD listens to their users and gives the
    users what they want, and right now they're not hearing any demand
    for
    AVX512.

    Personally, I call BS on that one. I can't imagine anyone in the HPC
    community saying "we'd like processors that offer only 1/2 the
    floating
    point performance of Intel processors". Sure, AMD can offer more
    cores,
    but with only AVX2, you'd need twice as many cores as Intel
    processors,
    all other things being equal.

    Last fall I evaluated potential new cluster nodes for a large cluster
    purchase using the HPL benchmark. I compared a server with dual
    AMD EPYC
    7H12 processors (128) cores to a server with quad Intel Xeon 8268
    processors (96 cores). I measured 5,389 GFLOPS for the Xeon 8268, and
    only 3,446.00 GFLOPS for the AMD 7H12. That's LINPACK score that only
    64% of the Xeon 8268 system, despite having 33% more cores.

     From what I've heard, the AMD processors run much hotter than the
    Intel
    processors, too, so I imagine a FLOPS/Watt comparison would be
    even less
    favorable to AMD.

    An argument can be made that for calculations that lend themselves to
    vectorization should be done on GPUs, instead of the main
    processors but
    the last time I checked, GPU jobs are still memory is limited, and
    moving data in and out of GPU memory can still take time, so I can
    see
    situations where for large amounts of data using CPUs would be
    preferred
    over GPUs.

    Your thoughts?

-- Prentice


AMD has studied this quite a bit in DOE's FastForward-2 and PathForward. I think Carlos' comment is on track. Having a unit that cannot be fed data quick enough is pointless. It is application dependent. If your working set fits in cache, then the vector units work well. If not, you have to move data which stalls compute pipelines. NERSC saw only a 10% increase in performance when moving from low core count Xeon CPUs with AVX2 to Knights Landing with many cores and AVX-512 when it should have seen an order of magnitude increase. Although Knights Landing had MCDRAM (Micron's not-quite HBM), other constraints limited performance (e.g., lack of enough memory references in flight, coherence traffic).

Fujitsu's ARM64 chip with 512b SVE in Fugaku does much better than Xeon with AVX-512 (or Knights Landing) because of the High Bandwidth Memory (HBM) attached and I assume a larger number of memory references in flight. The downside is the lack of memory capacity (only 32 GB per node). This shows that it is possible to get more performance with a CPU with a 512b vector engine. That said, it is not clear that even this CPU design can extract the most from the memory bandwidth. If you look at the increase in memory bandwidth from Summit to Fugaku, one would expect performance on real apps to increase by that amount as well. From the presentations that I have seen, that is not always the case. For some apps, the GPU architecture, with its coherence on demand rather than with every operation, can extract more performance.

AMD will add 512b vectors if/when it makes sense on real apps.
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
https://beowulf.org/cgi-bin/mailman/listinfo/beowulf

Reply via email to