[Beowulf] SATA II - PXE+NFS - diskless compute nodes

Michael Will mwill at penguincomputing.com
Fri Dec 8 13:14:31 PST 2006


Geoff Jacobs wrote:
> Mark Hahn wrote:
>   
>> it's interesting that SAS advertising has obscured the fact that SAS is
>> just a further development of SCSI, and not interchangable
>> with SATA.  for instance, no SATA controller will support any SAS disk,
>> and any SAS setup uses a form of encapsulation to communicate with
>> the foreign SATA protocol.  SAS disks follow the traditional price
>> formula of SCSI disks (at least 4x more than non-boutique disks),
>> and I suspect the rest of SAS infrastructure will be in line with that.
>>     
> Yes, SAS encapsulates SATA, but not vice-versa. The ability to use a
> hardware raid SAS controller with large numbers of inexpensive SATA
> drives is very attractive. I was also trying to be thorough.
>
>   
>>> and be mindful of reliability issues with desktop drives.
>>>       
>> I would claim that this is basically irrelevant for beowulf.
>> for small clusters (say, < 100 nodes), you'll be hitting a negligable
>> number of failures per year.  for larger clusters, you can't afford
>> any non-ephemeral install on the disks anyway - reboot-with-reimage
>> should only take a couple minutes more than a "normal" reboot.
>> and if you take the no-install (NFS root) approach (which I strongly
>> recommend) the status of a node-local disks can be just a minor node
>> property to be handled by the scheduler.
>>     
> PXE/NFS is absolutely the slickest way to go, but any service nodes
> should have some guarantee of reliability. In my experience, disks
> (along with power supplies) are two of the most common points of failure
Most of the clusters we configure for our customers use diskless compute 
nodes to minimize compute node failure for
precisely the reason you mentioned unless either the application can 
benefit from additional
local scratchspace (i.e. software raid0 over four sata drives allows to 
read/write large datastreams
at 280MB/s in a 1U server with 3TB of disk space on each compute node), 
or because they need to
sometimes run jobs that require more virtual memory than they can afford 
to put in physically -> local swapspace.

We find that customers don't typically want to pay for the premium for 
redundant power supplies+pdus+cabling
for the compute nodes through, that's something that is typically 
requested for head nodes and NFS servers.

Also we find that NFS-offloading on the NFS-server with the rapidfile 
card helps avoid scalability issues
where the NFS server bogs down under massively parallel requests from 
say 128 cores in a 32 compute node dual
cpu dual core cluster. The rapidfile card is a pci-x card with two fibre 
channel ports + two gige ports +
nfs/cifs offloading processor on the same card. Since most bulk data 
transfer is redirected from fibre channel to gige nfs
clients without passing through the NFS server cpu+ram itself, the nfs 
servers cpu load is not becoming the bottleneck,
we find it's rather the amount of spindles before saturating the two 
gige ports.

We configure clusters for our customers with Scyld Beowulf which does 
not nfs-mount
root but rather just nfs-mounts the home directories because of its 
particular lightweight
compute node model, (PXE booting into RAM) and so  does not run into the 
typical
nfs-root scalability issues.

Michael

Michael Will
SE Technical Lead / Penguin Computing / www.penguincomputing.com



More information about the Beowulf mailing list