Thanks Josh. Am I familiar with modifying Python code and PBS hook scripts? Yes - I have had my head under the hood of PBS hooks for a long time. Hence the pronounced stutter and my predelection to randomly scream out loud in public places.
On 16 June 2017 at 15:48, Josh Catana <jcat...@gmail.com> wrote: > I know they have a canned scheduler hook to run docker. If you're familiar > with python modifying their code to run singularity shouldn't be difficult. > I rewrote their hook to operate in my environment pretty easily. > > On Jun 16, 2017 4:29 AM, "John Hearns" <hear...@googlemail.com> wrote: > >> Lance, thankyou very much for the reply. I will look at Docker for those >> 'system' type tasks also. >> >> Regarding Singularity does anyone know much about Singularity integration >> with PBSPro? >> I guess I could actually ask Altair.... >> >> >> On 16 June 2017 at 01:30, Lance Wilson <lance.wil...@monash.edu> wrote: >> >>> Hi John, >>> In regards to your Singularity question we are using cgroups for the >>> containers. Mostly the containers are used in Slurm jobs which creates the >>> appropriate cgroups. We are also using the gpu driver passthrough >>> functionality of Singularity now for our machine learning and cryoem >>> processing containers which have the cgroups applied to gpus. >>> >>> Back to your systems containers questions many of our systems have been >>> put into docker containers as they run on same/similar operating system and >>> still need root to function correctly. Pretty much every new system thing >>> we do is scripted and put into a container so that we can recover quickly >>> in an outage scenario and move around things as part of our larger cloud >>> (private and public) strategy. >>> >>> Cheers, >>> >>> Lance >>> -- >>> Dr Lance Wilson >>> Senior HPC Consultant >>> Ph: 03 99055942 (+61 3 99055942 <+61%203%209905%205942> >>> Mobile: 0437414123 (+61 4 3741 4123) >>> Multi-modal Australian ScienceS Imaging and Visualisation Environment >>> (www.massive.org.au) >>> Monash University >>> >>> On 15 June 2017 at 20:06, John Hearns <hear...@googlemail.com> wrote: >>> >>>> I'm not sure this post is going to make a lot of sense. But please bear >>>> with me! >>>> For applications containers are possible using Singularity or Docker of >>>> course. >>>> >>>> In HPC clusters we tend to have several 'service node' activities, such >>>> as the cluster management/ head node, perhaps separate provisioning nodes >>>> to spread the load, batch queue system masters, monitoring setups, job >>>> submission and dedicated storage nodes. >>>> >>>> These can all of course be run on a single cluster head node in a small >>>> setup (with the exception of the storage nodes). In a larger setup you can >>>> run these services in virtual machines. >>>> >>>> What I am asking is anyone using technologies such as LXD containers to >>>> run these services? >>>> I was inspired by an Openstack talk by James Page at Canonical, where >>>> all the Opestack services were deployed by Juju charms onto LXD containers. >>>> So we pack all the services into containers on physical server(s) which >>>> makes moving them or re-deploying things very flexible. >>>> https://www.youtube.com/watch?v=5orzBITR3X8 >>>> >>>> While I'm talking abotu containers, is anyone deploying singularity >>>> containers in cgroups, and limiting the resources they can use (I'm >>>> specifically thinking of RDMA here). >>>> >>>> >>>> >>>> ps. I have a terrible sense of deja vu here... I think I asked the >>>> Singularity question a month ago. >>>> I plead insanity m'lord >>>> >>>> >>>> >>>> _______________________________________________ >>>> Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin >>>> Computing >>>> To change your subscription (digest mode or unsubscribe) visit >>>> http://www.beowulf.org/mailman/listinfo/beowulf >>>> >>>> >>> >> >> _______________________________________________ >> Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing >> To change your subscription (digest mode or unsubscribe) visit >> http://www.beowulf.org/mailman/listinfo/beowulf >> >>
_______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf