<div dir="ltr">Brice, thankyou for the reply. You have the answer - these systems have two DIMMS per socket, channels 0 and, so all channels are not populated.<div><br></div><div>I had the lstopo output and the tarball all ready for the OpenMPI list too! Shoudl have sent it over there.</div><div><br></div></div><div class="gmail_extra"><br><div class="gmail_quote">On 18 December 2016 at 21:28, Brice Goglin <span dir="ltr"><<a href="mailto:brice.goglin@gmail.com" target="_blank">brice.goglin@gmail.com</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">Hello<br>
Do you know if all your CPU memory channels are populated? CoD requires<br>
that each half of the CPU has some memory DIMMs (so that each NUMA node<br>
actually contains some memory). If both channels of one half are empty,<br>
the NUMA node might somehow disappear.<br>
<span class="HOEnZb"><font color="#888888">Brice<br>
</font></span><div class="HOEnZb"><div class="h5"><br>
<br>
<br>
<br>
Le 16/12/2016 23:26, Elken, Tom a écrit :<br>
> Hi John and Greg,<br>
><br>
> You showed Nodes 0 & 2 (no node 1) and a strange CPU assignment to nodes!<br>
> Even though you had Cluster On Die (CoD) Endabled in your BIOS, I have never seen that arrangement of Numa nodes and CPUs. You may have a bug in your BIOS or OS ?<br>
> With CoD enabled, I would have expected 4 NUMA nodes, 0-3, and 6 cores assigned to each one.<br>
><br>
> The Omni-Path Performance Tuning User Guide<br>
> <a href="http://www.intel.com/content/dam/support/us/en/documents/network-and-i-o/fabric-products/Intel_OP_Performance_Tuning_UG_H93143_v6_0.pdf" rel="noreferrer" target="_blank">http://www.intel.com/content/<wbr>dam/support/us/en/documents/<wbr>network-and-i-o/fabric-<wbr>products/Intel_OP_Performance_<wbr>Tuning_UG_H93143_v6_0.pdf</a><br>
> does recommend Disabling CoD in Xeon BIOSes (Table 2 on P. 12), but it's not considered a hard prohibition.<br>
> Disabling improves some fabric performance benchmarks, but Enabling helps some single-node applications performance, which could outweigh the fabric performance aspects.<br>
><br>
> -Tom<br>
><br>
>> -----Original Message-----<br>
>> From: Beowulf [mailto:<a href="mailto:beowulf-bounces@beowulf.org">beowulf-bounces@<wbr>beowulf.org</a>] On Behalf Of Greg<br>
>> Lindahl<br>
>> Sent: Friday, December 16, 2016 2:00 PM<br>
>> To: John Hearns<br>
>> Cc: Beowulf Mailing List<br>
>> Subject: Re: [Beowulf] NUMA zone weirdness<br>
>><br>
>> Wow, that's pretty obscure!<br>
>><br>
>> I'd recommend reporting it to Intel so that they can add it to the<br>
>> descendants of ipath_checkout / ipath_debug. It's exactly the kind of<br>
>> hidden gotcha that leads to unhappy systems!<br>
>><br>
>> -- greg<br>
>><br>
>> On Fri, Dec 16, 2016 at 03:52:34PM +0000, John Hearns wrote:<br>
>>> Problem solved.<br>
>>> I have changed the QPI Snoop Mode on these servers from<br>
>>> ClusterOnDIe Enabled to Disabled and they display what I take to be correct<br>
>>> behaviour - ie<br>
>>><br>
>>> [root@comp006 ~]# numactl --hardware<br>
>>> available: 2 nodes (0-1)<br>
>>> node 0 cpus: 0 1 2 3 4 5 6 7 8 9 10 11<br>
>>> node 0 size: 32673 MB<br>
>>> node 0 free: 31541 MB<br>
>>> node 1 cpus: 12 13 14 15 16 17 18 19 20 21 22 23<br>
>>> node 1 size: 32768 MB<br>
>>> node 1 free: 31860 MB<br>
>>> node distances:<br>
>>> node 0 1<br>
>>> 0: 10 21<br>
>>> 1: 21 10<br>
>> ______________________________<wbr>_________________<br>
>> Beowulf mailing list, <a href="mailto:Beowulf@beowulf.org">Beowulf@beowulf.org</a> sponsored by Penguin Computing<br>
>> To change your subscription (digest mode or unsubscribe) visit<br>
>> <a href="http://www.beowulf.org/mailman/listinfo/beowulf" rel="noreferrer" target="_blank">http://www.beowulf.org/<wbr>mailman/listinfo/beowulf</a><br>
> ______________________________<wbr>_________________<br>
> Beowulf mailing list, <a href="mailto:Beowulf@beowulf.org">Beowulf@beowulf.org</a> sponsored by Penguin Computing<br>
> To change your subscription (digest mode or unsubscribe) visit <a href="http://www.beowulf.org/mailman/listinfo/beowulf" rel="noreferrer" target="_blank">http://www.beowulf.org/<wbr>mailman/listinfo/beowulf</a><br>
<br>
______________________________<wbr>_________________<br>
Beowulf mailing list, <a href="mailto:Beowulf@beowulf.org">Beowulf@beowulf.org</a> sponsored by Penguin Computing<br>
To change your subscription (digest mode or unsubscribe) visit <a href="http://www.beowulf.org/mailman/listinfo/beowulf" rel="noreferrer" target="_blank">http://www.beowulf.org/<wbr>mailman/listinfo/beowulf</a><br>
</div></div></blockquote></div><br></div>