[Beowulf] HPC fault tolerance using virtualization

Egan Ford egan at sense.net
Tue Jun 16 10:05:17 PDT 2009


Ha!  :-)

I've put a few GigE systems in the Top100, and if the stars align you'll see
a Top20 GigE system in next weeks list.  That's ONE GigE to each node
oversubscribed 4:1.  Sadly no flashing lights, and since its 100% water
cooled with low velocity fans, there is almost no noise.

On Tue, Jun 16, 2009 at 10:23 AM, John Hearns <hearnsj at googlemail.com>wrote:

>
>
> 2009/6/16 Egan Ford <egan at sense.net>
>
>> I have no idea the state of VMs on IB.  That can be an issue with MPI.
>> Believe it or not, but most HPC sites do not use MPI.  They are all batch
>> systems where storage I/O is the bottleneck.
>
>
> Burn the Witch! Burn the Witch!
>
> Any HPC installation, if you want to show it off to alumni, august
> committees from grant awarding bodies etc.  and not get sand kicked in your
> face from the big boys in the Top 500 NEEDS an expensive infrastructure of
> various MPI libraries. Big, big switches with lots of flashing lights.
> Highly paid, pampered systems admins who must be treated like expensive
> racehorses, and not exercised too much every day. They need cool beers on
> tap and luxurious offices to relax in while they prepare to do that vital
> half hours work per day which keeps your Supercomputer flashing away and
> making noises.
>
>
> _______________________________________________
> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://www.scyld.com/pipermail/beowulf/attachments/20090616/fa6d9a64/attachment.html


More information about the Beowulf mailing list