[Beowulf] HPC fault tolerance using virtualization
egan at sense.net
Tue Jun 16 10:05:17 PDT 2009
I've put a few GigE systems in the Top100, and if the stars align you'll see
a Top20 GigE system in next weeks list. That's ONE GigE to each node
oversubscribed 4:1. Sadly no flashing lights, and since its 100% water
cooled with low velocity fans, there is almost no noise.
On Tue, Jun 16, 2009 at 10:23 AM, John Hearns <hearnsj at googlemail.com>wrote:
> 2009/6/16 Egan Ford <egan at sense.net>
>> I have no idea the state of VMs on IB. That can be an issue with MPI.
>> Believe it or not, but most HPC sites do not use MPI. They are all batch
>> systems where storage I/O is the bottleneck.
> Burn the Witch! Burn the Witch!
> Any HPC installation, if you want to show it off to alumni, august
> committees from grant awarding bodies etc. and not get sand kicked in your
> face from the big boys in the Top 500 NEEDS an expensive infrastructure of
> various MPI libraries. Big, big switches with lots of flashing lights.
> Highly paid, pampered systems admins who must be treated like expensive
> racehorses, and not exercised too much every day. They need cool beers on
> tap and luxurious offices to relax in while they prepare to do that vital
> half hours work per day which keeps your Supercomputer flashing away and
> making noises.
> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
> To change your subscription (digest mode or unsubscribe) visit
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the Beowulf