> And the OS-Link that connected the CPUs in Transputer land lives on as > SpaceWire, which is a fairly popular high speed serial interconnect for > spacecraft systems from 2 to 400 Mbps. (there is a faster version called > SpaceFibre for both copper and optical that does multi Gbps).
Never new that! Thanks for that bit of trivia. -- Doug > > On 8/23/19, 7:45 AM, "Beowulf on behalf of Douglas Eadline" > <beowulf-boun...@beowulf.org on behalf of deadl...@eadline.org> wrote: > > > > > Hi John > > > > No doom and gloom. > > > > It's in a purpose built workshop/computer room that I have; 42U > Rack, > cross draft cooling which is sufficient and 32AMP Power into the > PDUâÂÂs. The equipment is housed in the 42U Rack along with a > variety of > other machines such as Sun Enterprise 4000 and a 30 CPU Transputer > Cluster. None of it runs 24/7 and not all of it is on at the same > time, > mainly because of the cost of power :-/ > > I'm jealous. I recall recycling about 16 Transputers years ago > because they were on an ISA interface card (systems I did not > have anymore) Transputers were fun. > > And, I'm curious how many people have 42 inch racks at home. > Mine is in a purpose built closet in my office. My neighbor > is convinced I secretly work for the government. I also > have all kinds of hardware and a big 3D printer in my office, > which he says further confirms his assertion. My reply > "I'm just a computer geek." > > -- > Doug > > > > > Yeah the Tesla 1070âÂÂs scream like a bansheeâ¦.. > > > > I am planning on running it as power on, on demand setup, which I > already do through some HP iLo and APC PDU Scripts that I have for > these > machines. Until recently I have been running some of them as a > vSphere > cluster and others as standalone CUDA machines. > > > > So thatâÂÂs one vote for OpenHPC. > > > > Cheers > > > > Richard > > > >> On 21 Aug 2019, at 3:45 pm, John Hearns via Beowulf > >> <beowulf@beowulf.org> wrote: > >> > >> Add up the power consumption for each of those servers. If you plan > on > installing this in a domestic house or indeed in a normal office > environment you probably wont have enough amperage in the circuit you > intend to power it from. > >> Sorry to be all doom and gloom. > >> Also this setup will make a great deal of noise. If in a domestic > setting put it in the garage. > >> In an office setting the obvious place is a comms room but be > careful > about the ventilation. > >> Office comms rooms often have a single wall mounted air > conditioning > unit. Make SURE to run a temperature shutdown script. > >> This air con unit WILL fail over a weekend. > >> > >> Regarding the software stack I would look at OpenHPC. But that's > just me. > >> > >> > >> > >> > >> > >> On Wed, 21 Aug 2019 at 06:09, Dmitri Chubarov > <dmitri.chuba...@gmail.com <mailto:dmitri.chuba...@gmail.com>> wrote: > >> Hi, > >> this is a very old hardware and you would have to stay with a very > outdated software stack as 1070 cards are not supported by the recent > versions of NVIDIA Drivers and old versions of NVIDIA drivers do not > play well with modern kernels and modern system libraries.Unless you > are doing this for digital preservation, consider dropping 1070s out > of > the equation. > >> > >> Dmitri > >> > >> > >> On Wed, 21 Aug 2019 at 06:46, Richard Edwards <e...@fastmail.fm > >> <mailto:e...@fastmail.fm>> wrote: > >> Hi Folks > >> > >> So about to build a new personal GPU enabled cluster and am looking > for > peoples thoughts on distribution and management tools. > >> > >> Hardware that I have available for the build > >> - HP Proliant DL380/360 - mix of G5/G6 > >> - HP Proliant SL6500 with 8 GPU > >> - HP Proliant DL580 - G7 + 2x K20x GPU > >> -3x Nvidia Tesla 1070 (4 GPU per unit) > >> > >> Appreciate people insights/thoughts > >> > >> Regards > >> > >> Richard > >> _______________________________________________ > >> Beowulf mailing list, Beowulf@beowulf.org > <mailto:Beowulf@beowulf.org> > sponsored by Penguin Computing > >> To change your subscription (digest mode or unsubscribe) visit > >> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf > >> <https://beowulf.org/cgi-bin/mailman/listinfo/beowulf> > >> _______________________________________________ > >> Beowulf mailing list, Beowulf@beowulf.org > <mailto:Beowulf@beowulf.org> > sponsored by Penguin Computing > >> To change your subscription (digest mode or unsubscribe) visit > >> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf > >> <https://beowulf.org/cgi-bin/mailman/listinfo/beowulf> > >> _______________________________________________ > >> Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin > Computing To change your subscription (digest mode or unsubscribe) > visit > >> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf > > > > _______________________________________________ > > Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin > Computing > To change your subscription (digest mode or unsubscribe) visit > > https://beowulf.org/cgi-bin/mailman/listinfo/beowulf > > > > > -- > Doug > > > > -- > Doug > > _______________________________________________ > Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin > Computing > To change your subscription (digest mode or unsubscribe) visit > https://beowulf.org/cgi-bin/mailman/listinfo/beowulf > > > -- Doug _______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit https://beowulf.org/cgi-bin/mailman/listinfo/beowulf