<div dir="ltr">Yes, you belong. :-)</div><br><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Sat, Feb 23, 2019 at 9:41 AM Will Dennis <<a href="mailto:wdennis@nec-labs.com">wdennis@nec-labs.com</a>> wrote:<br></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left-width:1px;border-left-style:solid;border-left-color:rgb(204,204,204);padding-left:1ex">
<div lang="EN-US">
<div class="gmail-m_-5360851326755508208WordSection1">
<p class="MsoNormal">Hi folks,<u></u><u></u></p>
<p class="MsoNormal"><u></u> <u></u></p>
<p class="MsoNormal">I thought I’d give a brief introduction, and see if this list is a good fit for my questions that I have about my HPC-“ish” infrastructure...<u></u><u></u></p>
<p class="MsoNormal"><u></u> <u></u></p>
<p class="MsoNormal">I am a ~30yr sysadmin (“jack-of-all-trades” type), completely self-taught (B.A. is in English, that’s why I’m a sysadmin :-P) and have ended up working at an industrial research lab for a large multi-national IT company (<a href="http://www.nec-labs.com" target="_blank">http://www.nec-labs.com</a>).
In our lab we have many research groups (as detailed on the aforementioned website) and a few of them are now using “HPC” technologies like Slurm, and I’ve become the lead admin for these groups. Having no prior background in this realm, I’m learning as fast
as I can go :) <u></u><u></u></p>
<p class="MsoNormal"><u></u> <u></u></p>
<p class="MsoNormal">Our “clusters” are collections of 5-30 servers, all collections bought over years and therefore heterogeneous hardware, all with locally-installed OS (i.e. not trad head-node with PXE-booted diskless minions) which is as carefully controlled
as I can make it via standard OS install via Cobbler templates, and then further configured via config management (we use Ansible.) Networking is basic 10GbE between nodes (we do have Infiniband availability on one cluster, but it’s fell into disuse now since
the project that has required it has ended.) Storage is one or more traditional NFS servers (some use ZFS, some not.) We have within the past few years adopted Slurm WLM for a job-scheduling system on top of these collections, and now are up to three different
Slurm clusters, with I believe a fourth on the way.<u></u><u></u></p>
<p class="MsoNormal"><u></u> <u></u></p>
<p class="MsoNormal">My first question for this list is basically “do I belong here?” I feel there’s a lot of HPC concepts it would be good for me to learn, so as I can improve the various research group’s computing environments, but not sure if this list is
for much larger “true HPC” environments, or would be a good fit for a “HPC n00b” like me...<u></u><u></u></p>
<p class="MsoNormal"><u></u> <u></u></p>
<p class="MsoNormal">Thanks for reading, and let me know your opinions :)<u></u><u></u></p>
<p class="MsoNormal"><u></u> <u></u></p>
<p class="MsoNormal">Best,<u></u><u></u></p>
<p class="MsoNormal">Will<u></u><u></u></p>
</div>
</div>
_______________________________________________<br>
Beowulf mailing list, <a href="mailto:Beowulf@beowulf.org" target="_blank">Beowulf@beowulf.org</a> sponsored by Penguin Computing<br>
To change your subscription (digest mode or unsubscribe) visit <a href="http://www.beowulf.org/mailman/listinfo/beowulf" rel="noreferrer" target="_blank">http://www.beowulf.org/mailman/listinfo/beowulf</a><br>
</blockquote></div>