[Beowulf] [EXTERNAL] Re: Build Recommendations - Private Cluster
Douglas Eadline
deadline at eadline.org
Fri Aug 23 11:56:44 PDT 2019
> And the OS-Link that connected the CPUs in Transputer land lives on as
> SpaceWire, which is a fairly popular high speed serial interconnect for
> spacecraft systems from 2 to 400 Mbps. (there is a faster version called
> SpaceFibre for both copper and optical that does multi Gbps).
Never new that! Thanks for that bit of trivia.
--
Doug
>
> On 8/23/19, 7:45 AM, "Beowulf on behalf of Douglas Eadline"
> <beowulf-bounces at beowulf.org on behalf of deadline at eadline.org> wrote:
>
>
>
> > Hi John
> >
> > No doom and gloom.
> >
> > It's in a purpose built workshop/computer room that I have; 42U
> Rack,
> cross draft cooling which is sufficient and 32AMP Power into the
> PDUâÂÂs. The equipment is housed in the 42U Rack along with a
> variety of
> other machines such as Sun Enterprise 4000 and a 30 CPU Transputer
> Cluster. None of it runs 24/7 and not all of it is on at the same
> time,
> mainly because of the cost of power :-/
>
> I'm jealous. I recall recycling about 16 Transputers years ago
> because they were on an ISA interface card (systems I did not
> have anymore) Transputers were fun.
>
> And, I'm curious how many people have 42 inch racks at home.
> Mine is in a purpose built closet in my office. My neighbor
> is convinced I secretly work for the government. I also
> have all kinds of hardware and a big 3D printer in my office,
> which he says further confirms his assertion. My reply
> "I'm just a computer geek."
>
> --
> Doug
>
> >
> > Yeah the Tesla 1070âÂÂs scream like a bansheeâ¦..
> >
> > I am planning on running it as power on, on demand setup, which I
> already do through some HP iLo and APC PDU Scripts that I have for
> these
> machines. Until recently I have been running some of them as a
> vSphere
> cluster and others as standalone CUDA machines.
> >
> > So thatâÂÂs one vote for OpenHPC.
> >
> > Cheers
> >
> > Richard
> >
> >> On 21 Aug 2019, at 3:45 pm, John Hearns via Beowulf
> >> <beowulf at beowulf.org> wrote:
> >>
> >> Add up the power consumption for each of those servers. If you plan
> on
> installing this in a domestic house or indeed in a normal office
> environment you probably wont have enough amperage in the circuit you
> intend to power it from.
> >> Sorry to be all doom and gloom.
> >> Also this setup will make a great deal of noise. If in a domestic
> setting put it in the garage.
> >> In an office setting the obvious place is a comms room but be
> careful
> about the ventilation.
> >> Office comms rooms often have a single wall mounted air
> conditioning
> unit. Make SURE to run a temperature shutdown script.
> >> This air con unit WILL fail over a weekend.
> >>
> >> Regarding the software stack I would look at OpenHPC. But that's
> just me.
> >>
> >>
> >>
> >>
> >>
> >> On Wed, 21 Aug 2019 at 06:09, Dmitri Chubarov
> <dmitri.chubarov at gmail.com <mailto:dmitri.chubarov at gmail.com>> wrote:
> >> Hi,
> >> this is a very old hardware and you would have to stay with a very
> outdated software stack as 1070 cards are not supported by the recent
> versions of NVIDIA Drivers and old versions of NVIDIA drivers do not
> play well with modern kernels and modern system libraries.Unless you
> are doing this for digital preservation, consider dropping 1070s out
> of
> the equation.
> >>
> >> Dmitri
> >>
> >>
> >> On Wed, 21 Aug 2019 at 06:46, Richard Edwards <ejb at fastmail.fm
> >> <mailto:ejb at fastmail.fm>> wrote:
> >> Hi Folks
> >>
> >> So about to build a new personal GPU enabled cluster and am looking
> for
> peoples thoughts on distribution and management tools.
> >>
> >> Hardware that I have available for the build
> >> - HP Proliant DL380/360 - mix of G5/G6
> >> - HP Proliant SL6500 with 8 GPU
> >> - HP Proliant DL580 - G7 + 2x K20x GPU
> >> -3x Nvidia Tesla 1070 (4 GPU per unit)
> >>
> >> Appreciate people insights/thoughts
> >>
> >> Regards
> >>
> >> Richard
> >> _______________________________________________
> >> Beowulf mailing list, Beowulf at beowulf.org
> <mailto:Beowulf at beowulf.org>
> sponsored by Penguin Computing
> >> To change your subscription (digest mode or unsubscribe) visit
> >> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf
> >> <https://beowulf.org/cgi-bin/mailman/listinfo/beowulf>
> >> _______________________________________________
> >> Beowulf mailing list, Beowulf at beowulf.org
> <mailto:Beowulf at beowulf.org>
> sponsored by Penguin Computing
> >> To change your subscription (digest mode or unsubscribe) visit
> >> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf
> >> <https://beowulf.org/cgi-bin/mailman/listinfo/beowulf>
> >> _______________________________________________
> >> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin
> Computing To change your subscription (digest mode or unsubscribe)
> visit
> >> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf
> >
> > _______________________________________________
> > Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin
> Computing
> To change your subscription (digest mode or unsubscribe) visit
> > https://beowulf.org/cgi-bin/mailman/listinfo/beowulf
> >
>
>
> --
> Doug
>
>
>
> --
> Doug
>
> _______________________________________________
> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin
> Computing
> To change your subscription (digest mode or unsubscribe) visit
> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf
>
>
>
--
Doug
More information about the Beowulf
mailing list