[Beowulf] [gorelsky at stanford.edu: CCL:dual-core Opteron 275performance]
Joe Landman
landman at scalableinformatics.com
Wed Jul 13 09:56:09 PDT 2005
Hi Mikhail:
If you use numactl, you should have control over processor affinity
for a particular process. I am not sure how this ties in to MPI though,
so there may need to be some work there.
Joe
Mikhail Kuzminsky wrote:
> In message from Alan Louis Scheinine <scheinin at crs4.it> (Tue, 12 Jul
> 2005 12:24:27 +0200):
>
>> 1) Gerry Creager wrote "Hoowa!"
>> Since the results seem useful, I would like to add the following.
>> On dual-CPU boards with Athlon32 CPUs, the program "bolam" was
>> slow if
>> both CPUs on the board were used, it was better to have one MPICH
>> process
>> per compute node. This problem did not appear in another cluster
>> that had
>> Opteron dual-CPU boards (single-core), that is, two processes for
>> each node
>> did not cause a slowdown. This is an indication that "bolam" is at a
>> threshold for memory access being a bottleneck.
>
> The original post by S.Gorelsky (re-sent by E.Leitl) was about good
> scalability of 4cores/dual-CPUs Opteron 275 server on Gaussian 03
> DFT/test397 test. I'm testing just now like Supermicro server
> w/2*Opteron 275 but w/DDR333 instead of DDR400 used by S.Gorelsky.
> I used SuSE 9.0 w/2.4.21 kernel.
>
> I understood, that original results of S.Gorelsky were obtained, probably,
> for shared memory parallelization ! If I use G03 w/Linda (which
> is main parallelization tool for G03 - parallelization in shared
> memory model of G03 is available only for more restricted subset
> of quantum-chemical methods) - then the results are much more bad.
>
> On 4 cores I obtained speedup only 2.95 for Linda vs 3.6 for
> shared memory. The difference is, as I understand, simple because
> of data exchanges through RAM for the case of Linda; in shared memory
> model like memory traffic is absent.
> FYI: speedup by S.Gorelsky for 4 CPUs is 3.4 (hope that I calculated
> properly :-)).
>
> I also obtained similar results for other quantum-chemical methods which
> show that using of Linda/G03 may give bad scalability for
> dual-core Opteron.
> We also have some (developing by us) quantum-chemical application which
> is bandwidth-limited under parallelization, and using of 1 CPU (1 MPI
> process) per dual Xeon nodes for Myrinet/MPICH is strongly preferred. In
> the case of (dual single core CPUs)-Opteron nodes the situation is better.
>
> But now for 4cores/2CPUs per Opteron node to force the using of
> only 2 cores (from 4), by 1 for each chip, we'll need to have
> cpu affinity support in Linux.
>
> Yours
> Mikhail
>
>> A complication for this
>> interpretation is that the Athlon32 nodes use Linux kernel 2.4.21.
>> 2) Mikhail Kuzminsky asked "do you have "node interleave memory"
>> switched off?
>> Reading the BIOS:
>> Bank interleaving "Auto", there are two memory modules per CPU so
>> there
>> should be bank interleaving.
>> Node interleaving "Disable"
>> 3) In an email Guy Coates asked
>> > Did you need to use numa-tools to specify the CPU placement, or
>> did the
>> > kernel "do the right thing" by itself?
>> The kernel did the right thing by itself.
>> I have a question: what are numa-tools?
>> On the computer I find
>> man -k numa
>> numa (3) - NUMA policy library
>> numactl(8) - Control NUMA policy for processes or shared memory
>> rpm -qa | grep -i numa
>> numactl-0.6.4-1.13
>> Is numactl the "numa-tools"? Is there another package to consider
>> installing?
>> I see that numactl has many "man" pages.
>>
>> Reference, previous message:
>> >In all cases, 4 MPI processes on a machine with 4 cores (two
>> dual-core CPUs).
>> >Meteorology program 1, "bolam" CPU time, real time (in seconds)
>> > Linux kernel 2.6.9-11.ELsmp 122 128
>> > Linux kernel 2.6.12.2 64 77
>> >
>> >Meteorology program 2, "non-hydrostatic"
>> > Linux kernel 2.6.9-11.ELsmp 598 544
>> > Linux kernel 2.6.12.2 430 476
>>
>>
>> --
>>
>> Centro di Ricerca, Sviluppo e Studi Superiori in Sardegna
>> Center for Advanced Studies, Research, and Development in Sardinia
>>
>> Postal Address: | Physical Address for FedEx, UPS, DHL:
>> --------------- | -------------------------------------
>> Alan Scheinine | Alan Scheinine
>> c/o CRS4 | c/o CRS4
>> C.P. n. 25 | Loc. Pixina Manna Edificio 1
>> 09010 Pula (Cagliari), Italy | 09010 Pula (Cagliari), Italy
>>
>> Email: scheinin at crs4.it
>>
>> Phone: 070 9250 238 [+39 070 9250 238]
>> Fax: 070 9250 216 or 220 [+39 070 9250 216 or +39 070 9250 220]
>> Operator at reception: 070 9250 1 [+39 070 9250 1]
>> Mobile phone: 347 7990472 [+39 347 7990472]
>>
>
> _______________________________________________
> Beowulf mailing list, Beowulf at beowulf.org
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf
More information about the Beowulf
mailing list