[Beowulf] evaluating FLOPS capacity of our cluster
Tom Elken
tom.elken at qlogic.com
Mon May 11 13:52:18 PDT 2009
> On Behalf Of Rahul Nabar
>
> Rmax/Rpeak= 0.83 seems a good guess based on one very similar system
> on the Top500.
>
> Thus I come up with a number of around 1.34 TeraFLOPS for my cluster
> of 24 servers. Does the value seem reasonable ballpark? Nothing too
> accurate but I do not want to be an order of magnitude off. [maybe a
> decimal mistake in math! ]
You're in the right ballpark. I recently got 0.245 Tflops on HPL on a 4-node version of what you have (with Goto BLAS), so 6x that # is in the same ballpark as your 1.34 TF/s estimate. My CPUs were 2.3 GHz Opteron 2356 instead of your 2.2 GHz.
Greg is also right on the memory size being a factor allowing larger N to be used for HPL. I used a pretty small N on this HPL run since we were running it as part of a HPC Challenge suite run, and a smaller N can be better for PTRANS if you are interested in the non-HPL parts of HPCC (as I was).
> All 64 bit machines with a dual channel
> bonded Gigabit ethernet interconnect. AMD Quad-Core AMD Opteron(tm)
> Processor 2354.
As others have said, 50% is a more likely HPL efficiency for a large GigE cluster, but with your smallish cluster (24 nodes) and bonded channels, you would probably get closer to 80% than 50%.
-Tom
>
>
> PS. The Athelon was my typo, earlier sorry!
>
> --
> Rahul
> _______________________________________________
> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin
> Computing
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf
More information about the Beowulf
mailing list