[Beowulf] Scheduler question -- non-uniform memory allocation to MPI
unl at harvill.net
Thu Jul 30 08:34:44 PDT 2015
We run SLURM with cgroups for memory containment of jobs. When users
resources on our cluster many times they will specify the number of
(MPI) tasks and
memory per task. The reality of much of the software that runs is that
most of the
memory is used by MPI rank 0 and much less on slave processes. This is
and sometimes causes bad outcomes (OOMs and worse) during job runs.
AFAIK SLURM is not able to allow users to request a different amount of
for different processes in their MPI pool. We used to run Maui/Torque
and I'm fairly
certain that feature is not present in that scheduler either.
Does anyone know if any scheduler allows the user to request different
memory per process? We know we can move to whole-node assignment to remedy
this problem but there is resistance to that...
Holland Computing Center
More information about the Beowulf