Hi Dmitri, I have no specific application. I have done some CUDA Enabled OpenCV for realtime video stitching, pattern recognition etc in the past. Was planning on spending some time learning more about CUDA and getting into MPICH. I think the K20x’s might still be OK for tensor flow. However this exercise is for me more about infrastructure build, management and learning than a given application.
Would certainly be interested in insights to the s/w stack. Cheers Richard > On 21 Aug 2019, at 4:06 pm, Dmitri Chubarov <dmitri.chuba...@gmail.com> wrote: > > Hi Richard, > > I am speaking from experience of keeping up a small cluster of 4 Supermicro > boxes with a total of 16 C2070 cards. We had to freeze NVIDIA Driver updates > as Fermi cards are not supported by the latest drivers as well. This means > you can use CUDA but not the latest versions of NVIDIA SDK. Machine Learning > applications are out since later version of TensorFlow require later versions > of NVIDIA Drivers and SDK. So this cluster runs some computational chemistry > codes that are less demanding in terms of CUDA features. I can probably give > you details of the software stack off the list. > > What would be good to keep in the list thread is the information on the type > of applications that you intend to use the cluster for. > > > > On Wed, 21 Aug 2019 at 12:47, Richard Edwards <e...@fastmail.fm > <mailto:e...@fastmail.fm>> wrote: > Hi Dmitri > > Thanks for the response. > > Yes old hardware but as I said it is for a personal cluster. I have also put > M2070’s in one of the 1070 chases as they are basically 4 slot PCI > expansions. I have various other M2050/M2070/M2090/K20x cards around so > depending on time I can certainly get more bang than the C1060’s that are in > there now. I am prepared to live with the pain of older drivers, potentially > having to use older linux distributions and not being able to support much > beyond CUDA 2.0... > > Yes I could go out and purchase probably a couple of newer cards and get the > same performance or better but this is more about the exercise and the > learning. > > So maybe the hardware list was a distraction. What are people using as the > predominant distro and management tools? > > cheers > > Richard > > > >> On 21 Aug 2019, at 3:08 pm, Dmitri Chubarov <dmitri.chuba...@gmail.com >> <mailto:dmitri.chuba...@gmail.com>> wrote: >> >> Hi, >> this is a very old hardware and you would have to stay with a very outdated >> software stack as 1070 cards are not supported by the recent versions of >> NVIDIA Drivers and old versions of NVIDIA drivers do not play well with >> modern kernels and modern system libraries.Unless you are doing this for >> digital preservation, consider dropping 1070s out of the equation. >> >> Dmitri >> >> >> On Wed, 21 Aug 2019 at 06:46, Richard Edwards <e...@fastmail.fm >> <mailto:e...@fastmail.fm>> wrote: >> Hi Folks >> >> So about to build a new personal GPU enabled cluster and am looking for >> peoples thoughts on distribution and management tools. >> >> Hardware that I have available for the build >> - HP Proliant DL380/360 - mix of G5/G6 >> - HP Proliant SL6500 with 8 GPU >> - HP Proliant DL580 - G7 + 2x K20x GPU >> -3x Nvidia Tesla 1070 (4 GPU per unit) >> >> Appreciate people insights/thoughts >> >> Regards >> >> Richard >> _______________________________________________ >> Beowulf mailing list, Beowulf@beowulf.org <mailto:Beowulf@beowulf.org> >> sponsored by Penguin Computing >> To change your subscription (digest mode or unsubscribe) visit >> https://beowulf.org/cgi-bin/mailman/listinfo/beowulf >> <https://beowulf.org/cgi-bin/mailman/listinfo/beowulf> >
_______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit https://beowulf.org/cgi-bin/mailman/listinfo/beowulf