[Beowulf] Problems with Dell M620 and CPU power throttling
Don Holmgren
djholm at fnal.gov
Fri Aug 30 08:23:21 PDT 2013
It might be worth fooling a bit with the cpufreq settings, down in
/sys/devices/system/cpu/cpuX/cpufreq
(where X=cpu#, one per core) To prevent non-thermal throttling, you can do
for each core
echo userspace > scaling_governor
cat scaling_max_freq
echo 2400000 > scaling_setspeed
(where substitute the max_freq reported for the 2400000). For this to work you
need the specific cpufreq driver for your processor loaded. For our (non-Dell)
SB servers it's acpi_cpufreq. In RedHat, the cpuspeed service loads the
relevent drivers, not sure if there is a similar service in other distros.
The above will lock the the cores at the max_freq, although if they get too hot
they will still throttle down in speed. There are statistics available on
frequency changes from thermal throttling in
/sys/devices/system/cpu/cpu0/thermal_throttle/
although I haven't used them, so I'm not sure about their functionality.
If you do a
modprobe cpufreq_stats
then a new directory
/sys/devices/system/cpu/cpu0/cpufreq/stats
will show up that has statistics about cpu speed changes. I'm not sure whether
thermal throttling changes will also show here or not. On one of our large
Opteron clusters, we had a handful of nodes with somewhat similar slowdown
problems as you are seeing on your SB's. We now lock their frequencies, and we
monitor /sys/devices/system/cpu/cpu0/cpufreq/stats/total_trans (which give total
number of speed changes), alarming when total_trans is non-zero.
Don Holmgren
Fermilab
On Fri, 30 Aug 2013, Bill Wichser wrote:
> Since January, when we installed an M620 Sandybridge cluster from Dell,
> we have had issues with power and performance to compute nodes. Dell
> apparently continues to look into the problem but the usual responses
> have provided no solution. Firmware, BIOS, OS updates all are fruitless.
>
> The problem is that the node/CPU is power capped. We first detected
> this with the STREAM benchmark, a quick run, which shows memory
> bandwidth around 2000 instead of the normal 13000 MB/s. When the CPU is
> in the C0 state, this drops to around 600.
>
> The effect appears randomly across the entire cluster with 5-10% of the
> nodes demonstrating some slower performance. We don't know what
> triggers this. Using "turbostat" we can see that the GHz of the cores
> is >= 1 in most cases, dropping to about 0.2 in some of the worst cases.
> Looking at the power consumption by either the chassis GUI or using
> "impitool sdr list" we see that there is only about 80 watts being used.
>
> We run the RH 6.x release and are up to date with kernel/OS patches.
> All firmware is up to date. Chassis power is configured as
> non-redundant. tuned is set for performance. Turbo mode is
> on/hyperthreading is off/performance mode is set in BIOS.
>
> A reboot does not change this problem. But a power cycle returns the
> compute node to normal again. Again, we do not know what triggers this
> event. We are not overheating the nodes. But while applications are
> running, something triggers an event where this power capping takes effect.
>
> At this point we remain clueless about what is causing this to happen.
> We can detect the condition now and have been power cycling the nodes in
> order to reset.
>
> If anyone has a clue, or better yet, solved the issue, we'd love to hear
> the solution!
>
> Thanks,
> Bill
More information about the Beowulf
mailing list