[Beowulf] NUMA zone weirdness

John Hearns hearnsj at googlemail.com
Mon Dec 19 08:33:37 PST 2016


All, thankyou very much for looking at this and Happy Christmas when it
comes.
And a Guid New Year tae ye a.

Number of DIMMS doubled, all channels now populated.  Et voila:

(HT is on, COD is enabled)

[root at comp007 ~]# numactl --hardware
available: 4 nodes (0-3)
node 0 cpus: 0 1 2 3 4 5 24 25 26 27 28 29
node 0 size: 32673 MB
node 0 free: 31712 MB
node 1 cpus: 6 7 8 9 10 11 30 31 32 33 34 35
node 1 size: 32768 MB
node 1 free: 31926 MB
node 2 cpus: 12 13 14 15 16 17 36 37 38 39 40 41
node 2 size: 32768 MB
node 2 free: 31972 MB
node 3 cpus: 18 19 20 21 22 23 42 43 44 45 46 47
node 3 size: 32768 MB
node 3 free: 31953 MB
node distances:
node   0   1   2   3
  0:  10  11  21  21
  1:  11  10  21  21
  2:  21  21  10  11
  3:  21  21  11  10


[root at comp007 ~]# lscpu
Architecture:          x86_64
CPU op-mode(s):        32-bit, 64-bit
Byte Order:            Little Endian
CPU(s):                48
On-line CPU(s) list:   0-47
Thread(s) per core:    2
Core(s) per socket:    12
Socket(s):             2
NUMA node(s):          4
Vendor ID:             GenuineIntel
CPU family:            6
Model:                 79
Model name:            Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz
Stepping:              1
CPU MHz:               1200.289
BogoMIPS:              4396.61
Virtualization:        VT-x
L1d cache:             32K
L1i cache:             32K
L2 cache:              256K
L3 cache:              15360K
NUMA node0 CPU(s):     0-5,24-29
NUMA node1 CPU(s):     6-11,30-35
NUMA node2 CPU(s):     12-17,36-41
NUMA node3 CPU(s):     18-23,42-47





On 19 December 2016 at 03:26, John Hearns <hearnsj at googlemail.com> wrote:

> Brice, thankyou for the reply. You have the answer - these systems have
> two DIMMS per socket, channels 0 and, so all channels are not populated.
>
> I had the lstopo output and the tarball all ready for the OpenMPI list
> too! Shoudl have sent it over there.
>
>
> On 18 December 2016 at 21:28, Brice Goglin <brice.goglin at gmail.com> wrote:
>
>> Hello
>> Do you know if all your CPU memory channels are populated? CoD requires
>> that each half of the CPU has some memory DIMMs (so that each NUMA node
>> actually contains some memory). If both channels of one half are empty,
>> the NUMA node might somehow disappear.
>> Brice
>>
>>
>>
>>
>> Le 16/12/2016 23:26, Elken, Tom a écrit :
>> > Hi John and Greg,
>> >
>> > You showed  Nodes 0 & 2 (no node 1) and a strange CPU assignment to
>> nodes!
>> > Even though you had Cluster On Die (CoD) Endabled in your BIOS, I have
>> never seen that arrangement of Numa nodes and CPUs.  You may have a bug in
>> your BIOS or OS ?
>> > With CoD enabled, I would have expected 4 NUMA nodes, 0-3, and 6 cores
>> assigned to each one.
>> >
>> > The Omni-Path Performance Tuning User Guide
>> > http://www.intel.com/content/dam/support/us/en/documents/net
>> work-and-i-o/fabric-products/Intel_OP_Performance_Tuning_
>> UG_H93143_v6_0.pdf
>> > does recommend Disabling CoD in Xeon BIOSes  (Table 2 on P. 12), but
>> it's not considered a hard prohibition.
>> > Disabling improves some fabric performance benchmarks, but Enabling
>> helps some single-node applications performance, which could outweigh the
>> fabric performance aspects.
>> >
>> > -Tom
>> >
>> >> -----Original Message-----
>> >> From: Beowulf [mailto:beowulf-bounces at beowulf.org] On Behalf Of Greg
>> >> Lindahl
>> >> Sent: Friday, December 16, 2016 2:00 PM
>> >> To: John Hearns
>> >> Cc: Beowulf Mailing List
>> >> Subject: Re: [Beowulf] NUMA zone weirdness
>> >>
>> >> Wow, that's pretty obscure!
>> >>
>> >> I'd recommend reporting it to Intel so that they can add it to the
>> >> descendants of ipath_checkout / ipath_debug. It's exactly the kind of
>> >> hidden gotcha that leads to unhappy systems!
>> >>
>> >> -- greg
>> >>
>> >> On Fri, Dec 16, 2016 at 03:52:34PM +0000, John Hearns wrote:
>> >>> Problem solved.
>> >>> I have changed the QPI Snoop Mode on these servers from
>> >>> ClusterOnDIe Enabled to Disabled and they display what I take to be
>> correct
>> >>> behaviour - ie
>> >>>
>> >>> [root at comp006 ~]# numactl --hardware
>> >>> available: 2 nodes (0-1)
>> >>> node 0 cpus: 0 1 2 3 4 5 6 7 8 9 10 11
>> >>> node 0 size: 32673 MB
>> >>> node 0 free: 31541 MB
>> >>> node 1 cpus: 12 13 14 15 16 17 18 19 20 21 22 23
>> >>> node 1 size: 32768 MB
>> >>> node 1 free: 31860 MB
>> >>> node distances:
>> >>> node   0   1
>> >>>   0:  10  21
>> >>>   1:  21  10
>> >> _______________________________________________
>> >> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin
>> Computing
>> >> To change your subscription (digest mode or unsubscribe) visit
>> >> http://www.beowulf.org/mailman/listinfo/beowulf
>> > _______________________________________________
>> > Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin
>> Computing
>> > To change your subscription (digest mode or unsubscribe) visit
>> http://www.beowulf.org/mailman/listinfo/beowulf
>>
>> _______________________________________________
>> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
>> To change your subscription (digest mode or unsubscribe) visit
>> http://www.beowulf.org/mailman/listinfo/beowulf
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.beowulf.org/pipermail/beowulf/attachments/20161219/ddcdc4b1/attachment-0001.html>


More information about the Beowulf mailing list