[Beowulf] Re: Help: Building a disked Beowulf / Chaitanya Krishna, Brown et al
Ed Karns
edkarns at firewirestuff.com
Thu Aug 25 09:17:25 PDT 2005
On Aug 25, 2005, at 3:39 AM, beowulf-request at beowulf.org wrote:
>> ... the resources that I have are these:
>>
>> 1 Intel Pentium 4 3 Ghz Procs 10
>> 2 Intel Mother boards 10
>> 3 200 GB SATA Hard disks 10
>> 4 120 GB IDE Hard disks 10
>> 5 Network cards 10 + 1 (1 extra for master)
>> 6 Some already present switches
>>
>> All the IDE drives will be primary (the OS will reside on this) and
>> the SATA drives will be use as secondary drives for storage)
>>
>> My plan (and requirement) is the following:
>>
>> 1 To get the cluster up and running parallel jobs.
>> 2 The way I intend to do 1 is this. Install the OS (SuSE 9.3 Pro) on
>> the master and install bare-bones ( I am not sure, but may be
>> something
>> like kernel, NFS and/or NIS, SSH, etc) on the rest of the nodes so
>> that I am able to run parallel jobs as well as serial jobs on the
>> nodes. Will require help on this.
>>
>
> Your hardware looks perfectly reasonable for a small cluster.
> Let's hope that your NICs and switches "match" in some way --
> enough ports, gigabit ports and gigabit cards, whatever. One has
> to wonder a bit about why the nodes have both a 120 GB IDE and 200
> GB SATA drive instead of e.g. 2x[120,200] GB SATA only. I've never
> mixed drives like this and would expect that it works but would
> worry that it might do something to performance (Mark Hahn usually
> is the answer man as far as the overall IDE drive subsystem is
> concerned:-).
* Hardware suggestions:
Take the "master" off of any alternate network until complete
debugging of the cluster is accomplished (unplug it, at the least,
and remove that "alternate" NIC if possible) ... set the whole thing
up as a completely stand alone cluster until it works as required.
I would also wonder about the switches (10/100baseT = :>] or Gigabit
switches = :>) ?) If two switches, then "balance" the loads = same
number of CAT5 connections on each switch = 5 & 5 plus jumper, if
three switches = 3 & 3 & 4 plus jumpers.
All BIOS configurations on all systems should closely match,
especially I/O port configurations. All NIC (network cards) should
match = brand name and model type where possible.
Although I do not speak from authority on this type of x86 cluster,
my best guess to increase performance would be to use the SATA
drivers completely for the OS and cluster work and use the IDE drives
for mirroring and backup (the exact reverse of your
configurations) ... keeping the IDE drives off of the cluster if
possible ... or even remove the IDE drives from the systems and make
a RAID array for the cluster network.
web-pages
* OS and software configurations = Trust Mr. Hearns', DGS' and Mr.
Brown's suggestions.
From John Hearns:
" ... You could do worse than to consult Robert Brown's web-pages,
Google for Brahma and Duke University. Also get a copy of the OReilly
book on Linux clustering, the latest one. ..."
From DGS:
" ... You should look into some of the cluster "toolkits". Free ones
include OSCAR, ROCKS, Warewulf, and oneSIS. My favorite is Warewulf,
though ROCKS is probably the nearest you can get to a "cluster in a
box" for free. ..."
From Robert Brown:
" ... I'm assuming that the NICs are PXE-capable and that you've got
a KVM setup that you can move from machine to machine somehow to set
the BIOS and manage at least the initial install. ..."
Ed Karns
FireWireStuff.com
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.beowulf.org/pipermail/beowulf/attachments/20050825/670d50e5/attachment.html>
More information about the Beowulf
mailing list