[Beowulf] cluster deployment and config management
John Hearns
hearnsj at googlemail.com
Tue Sep 5 14:10:02 PDT 2017
Regarding Rocks clusters, permit me to vent a little.
In my last employ we provided Rocks clusters.... Rocks is firmly embedded
in the Redhat 6 era with out of date 2.6 kernels.
It uses kickstart for installations (which is OK). However with modern
generations of Intel processors you get a warning from Kickstart that the
architecture is unsupported... so you must add the flag to accept
unsupported architectures to your kickstart.cfg
If at this point the Forbidden Planet "Danger Will Robinson!" soundbite is
not playing on your head then I have a bridge to sell you in London...
Rocks version 7 is supposed to be on the horizon, but I doubt it. It may
have been launched, but I left the Rocks list five months ago and I do not
feel the lack. Sorry Rocks types if you are on here... no intention to give
you a kicking.
On 5 September 2017 at 20:19, psc <pscadmin at avalon.umaryland.edu> wrote:
> Hey everyone, .. any idea what happened with perceus?
> http://www.linux-mag.com/id/6386/
> https://github.com/perceus/perceus
>
> .. yeah; what happened with Arthur Stevens (Perceus, GravityFS/OS Green
> Provisioning, etc. ) where he is now; who is maintain, if anyone, perceus ?
>
> .. and come on Greg K. ... we know you are luring there somewhere being
> busy with singularity
> http://singularity.lbl.gov/ (kudos .. great job as always !!!)
> .. wasn't perceus yours original baby?
> https://gmkurtzer.github.io/
> .. can you bring some light what happened with the perceus project? .. I'd
> love to see it integrated with singularity -- that would made my
> day/month/year !!!!!!
>
> thanks!
> cheers,
> psc
>
> p.s. .. there there used to be rocks clusters (not sure about it's status
> these days)
> http://www.rocksclusters.org/wordpress/
>
> p.s.s. .. I'd say Warewulf is the "best" bet in most cases .. why keep
> reinventing the wheel ?
>
>
> On 09/05/2017 01:43 PM, beowulf-request at beowulf.org wrote:
>
>> Send Beowulf mailing list submissions to
>> beowulf at beowulf.org
>>
>> To subscribe or unsubscribe via the World Wide Web, visit
>> http://www.beowulf.org/mailman/listinfo/beowulf
>> or, via email, send a message with subject or body 'help' to
>> beowulf-request at beowulf.org
>>
>> You can reach the person managing the list at
>> beowulf-owner at beowulf.org
>>
>> When replying, please edit your Subject line so it is more specific
>> than "Re: Contents of Beowulf digest..."
>>
>>
>> Today's Topics:
>>
>> 1. Re: cluster deployment and config management (Joe Landman)
>> 2. Re: cluster deployment and config management (Arif Ali)
>> 3. RAID5 rebuild, remount with write without reboot? (mathog)
>> 4. Re: RAID5 rebuild, remount with write without reboot?
>> (John Hearns)
>>
>>
>> ----------------------------------------------------------------------
>>
>> Message: 1
>> Date: Tue, 5 Sep 2017 08:20:03 -0400
>> From: Joe Landman <joe.landman at gmail.com>
>> To: beowulf at beowulf.org
>> Subject: Re: [Beowulf] cluster deployment and config management
>> Message-ID: <2da2937a-9055-6514-3d24-a739aee12845 at gmail.com>
>> Content-Type: text/plain; charset=utf-8; format=flowed
>>
>>
>> Good morning ...
>>
>>
>> On 09/05/2017 01:24 AM, Stu Midgley wrote:
>>
>>> Morning everyone
>>>
>>> I am in the process of redeveloping our cluster deployment and config
>>> management environment and wondered what others are doing?
>>>
>>> First, everything we currently have is basically home-grown.
>>>
>> Nothing wrong with this, if it adequately solves the problem. Many of
>> the frameworks people use for these things are highly opinionated, and
>> often, you'll find their opinions grate on your expectations. At
>> $dayjob-1, I developed our own kit precisely because so many of the
>> other toolkits did little to big things wrong; not simply from an
>> opinion point of view, but actively made specific errors that the
>> developers glossed over as that aspect was unimportant to them ... while
>> being of critical importance to me and my customers at the time.
>>
>> Our cluster deployment is a system that I've developed over the years
>>> and is pretty simple - if you know BASH and how pxe booting works. It
>>> has everything from setting the correct parameters in the bios, zfs
>>> ram disks for the OS, lustre for state files (usually in /var) - all
>>> in the initrd.
>>>
>>> We use it to boot cluster nodes, lustre servers, misc servers and
>>> desktops.
>>>
>>> We basically treat everything like a cluster.
>>>
>> The most competent baked distro out there for this was (in the past,
>> haven't used it recently) Warewulf. See https://github.com/warewulf/ .
>> Still under active development, and Greg and team do a generally great
>> job. Least opinionated distro around, most flexible, and some of the
>> best tooling.
>>
>> However... we do have a proliferation of images... and all need to be
>>> kept up-to-date and managed. Most of the changes from one image to
>>> the next are config files.
>>>
>> Ahhh ... One of the things we did with our toolchain (it is open source,
>> I've just never pushed it to github) was to completely separate booting
>> from configuration. That is, units booted to an operational state
>> before we applied configuration. This was in part due to long
>> experience with nodes hanging during bootup with incorrect
>> configurations. If you minimize the chance for this, your nodes
>> (barring physical device failure) always boot. The only specific
>> opinion we had w.r.t. this system was that the nodes had to be bootable
>> via PXE, and there fore a working dhcp needed to exist on the network.
>>
>> Post boot configuration, we drove via a script that downloaded and
>> launched other scripts. Since we PXE booted, network addresses were
>> fine. We didn't even enforce final network address determination on PXE
>> startup.
>>
>> We looked at the booting process as a state machine. Lower level was
>> raw hardware, no power. Subsequent levels were bios POST, PXE of
>> kernel, configuration phase. During configuration phase *everything*
>> was on the table w.r.t. changes. We could (and did) alter networking,
>> using programmatic methods, databases, etc. to determine and configure
>> final network configs. Same for disks, and other resources.
>>
>> Configuration changes could be pushed post boot by updating a script and
>> either pushing (not normally recommended for clusters of reasonable
>> size) or triggering a pull/run cycle for that script/dependencies.
>>
>> This allowed us to update images and configuration asynchronously.
>>
>> We had to manage images, but this turned out to be generally simple. I
>> was in the midst of putting image mappings into a distributed object
>> store when the company died. Config store is similarly simple, again
>> using the same mechanisms, and could be driven entirely programmatically.
>>
>> Of course, for the chef/puppet/ansible/salt/cloudformation/... people,
>> we could drive their process as well.
>>
>>
>> We don't have a good config management (which might, hopefully, reduce
>>> the number of images we need). We tried puppet, but it seems everyone
>>> hates it. Its too complicated? Not the right tool?
>>>
>> Highly opinionated config management is IMO (and yes, I am aware this is
>> redundant humor) generally a bad idea. Config management that gets out
>> of your way until you need it is the right approach. Which is why we
>> never tried to dictate what config management our users would use. We
>> simply handled getting the system up to an operational state, and they
>> could use ours, theirs, or Frankensteinian kludges.
>>
>> I was thinking of using git for config files, dumping a list of rpm's,
>>> dumping the active services from systemd and somehow munging all that
>>> together in the initrd. ie. git checkout the server to get config
>>> files and systemctl enable/start the appropriate services etc.
>>>
>>> It started to get complicated.
>>>
>>> Any feedback/experiences appreciated. What works well? What doesn't?
>>>
>> IMO things that tie together config and booting are problematic at
>> scale. Leads to nearly unmanageable piles of images, as you've
>> experienced. Booting to an operational state, and applying all config
>> post boot (ask me about my fstab replacement some day), makes for a very
>> nice operational solution that scales wonderfully .... you can replicate
>> images to local image servers if you wish, replicate config servers,
>> load balance the whole thing to whatever scale you need.
>>
>>
>> Thanks.
>>>
>>>
>>>
>>> --
>>> Dr Stuart Midgley
>>> sdm900 at gmail.com <mailto:sdm900 at gmail.com>
>>>
>>>
>>> _______________________________________________
>>> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
>>> To change your subscription (digest mode or unsubscribe) visit
>>> http://www.beowulf.org/mailman/listinfo/beowulf
>>>
>>
>
> _______________________________________________
> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.beowulf.org/pipermail/beowulf/attachments/20170905/47959e01/attachment-0001.html>
More information about the Beowulf
mailing list