[Beowulf] Here we go again
bdobbins at gmail.com
Thu Dec 12 07:06:03 PST 2019
I've seen pretty decent performance on the AMD processors, and was even
told by AMD people to use the Intel compiler -- but when doing that, we
specify the processor type (eg, AVX capabilities), and it works pretty
well. However, I don't have any experience using the MKL on them.
That said, looking at the numbers, it's pretty interesting that there's
roughly a factor of 2 from the AVX2 (OpenBLAS) -> AVX512 (MKL) results on
Intel, and with the two systems being relatively comparable with OpenBLAS
(AVX2). Then it's *roughly* a factor of 8 going from the MKL on Intel to
the MKL on AMD, and since AVX512 is 8 x 64 floats, it seems it could just
be it's not using any vectorization whatsoever on AMD... presumably because
Intel claims they can't recognize the chip? That said, I'd love to see the
author try after setting:
That might be an easy fix, if it works.
Anyone got a Zen2 system with NumPy & the MKL to try it with?
On Thu, Dec 12, 2019 at 7:35 AM Douglas Eadline <deadline at eadline.org>
> Anyone see anything like this with Epyc, i.e. poor AMD performance
> when using Intel compilers or MKL?
> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
> To change your subscription (digest mode or unsubscribe) visit
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the Beowulf