[Beowulf] AMD 8 cores vs 12 cores CPUs and Infiniband

Mark Hahn hahn at mcmaster.ca
Tue Mar 29 21:02:33 PDT 2011

> I have been using single card on Magny-Cours with no issues at all. You can

interesting.  what adjustments have you made to the MPI stack to permit this?
we've had a variety of apps that fail intermittently on high-core nodes.
I have to say I was surprised such a thing came up - not sure whether it's 
inherent to IB or a result of the openmpi stack.  our usual way to test 
this is to gradually reduce the ranks-per-node for the job until it starts
to work.  an interesting cosmology code works at 1 pppn but not 3 ppn
on our recent 12c MC, mellanox QDR cluster.

> Using 16 cores per node (8 * 2) seem the 'safe' option, but the 24 cores
> (12 * 2) is better in term of price per job. Our CFD applications using MPI
> (OpenMPI) may need to do about 15 'MPI_allreduce' calls in one seccond or
> less, and we may probably using a pool of 1500 cores.

but will that allreduce be across 1500 cores?  I can get you a scaling curve
for the previously mentioned MC cluster (2.2 GHz).

> 2 - I've heard that QLogic behavior is better in terms of QP creation, I

well, they've often bragged about message rates - I'm not sure how relate
that is to QP creation.  I'd be interested to hear of some experiences, too.

regards, mark hahn.

More information about the Beowulf mailing list