We have our latest cluster software stack for a distributed set of clusters 
built on Ansible: 
https://github.com/CSC-IT-Center-for-Science/fgci-ansible 

A recent presentation at the SLURM User Group on Ansiblizing SLURM: 
https://gitpitch.com/CSC-IT-Center-for-Science/ansible-role-slurm/gitpitch 

I see benefits also in being able to share playbooks and collaborate on 
improving them with other teams in our organization and the Universities, even 
ones working in non-HPC areas. 

Best regards, 
Olli-Pekka 
-- 
Olli-Pekka Lehto 
Development Manager 
Computing Platforms 
CSC - IT Center for Science Ltd. 
E-Mail: olli-pekka.le...@csc.fi 
Tel: +358 50 381 8604 
skype: oplehto // twitter: ople 

> From: "Craig Andrew" <cband...@wi.mit.edu>
> To: "Tim Cutts" <t...@sanger.ac.uk>
> Cc: beowulf@beowulf.org
> Sent: Wednesday, 28 September, 2016 18:01:59
> Subject: Re: [Beowulf] more automatic building

> I agree with Tim.

> We are finishing up an Ansible install and it has worked well for us.

> Initially, we used it internally to help standardize our cluster builds, but 
> is
> has many more uses. We recently used it to provision a VM that we saved off 
> and
> uploaded to Amazon for building an AMI. You can also use it to change
> attributes on your running systems. I have used at Cobler in the past and it
> works well, too. I just find Ansible to be a little easier.

> Good luck,
> Craig

> Craig Andrew
> Manager of Systems Administration
> Whitehead Institute for Biomedical Research

> From: "Tim Cutts" <t...@sanger.ac.uk>
> To: "Mikhail Kuzminsky" <mikk...@mail.ru>, beowulf@beowulf.org
> Sent: Wednesday, September 28, 2016 10:46:41 AM
> Subject: Re: [Beowulf] more automatic building

> Any number of approaches will work. When I used to do this years ago (I've 
> long
> since passed on the technical side) I'd PXE boot, partition the hard disk and
> set up a provisioning network and base OS install using the Debian FAI (Fully
> Automated Install) system, and then use cfengine to configure the machine once
> it had come in that minimal state. This approach was used across the board for
> all of our Linux boxes, from Linux desktops to database servers to HPC compute
> nodes.

> These days the team uses tools like cobbler and ansible to achieve the same
> thing. There are lots of ways to do it, but the principle is the same.

> Tim

> --

> Head of Scientific Computing

> Wellcome Trust Sanger Institute

> On 28/09/2016, 15:34, "Beowulf on behalf of Mikhail Kuzminsky" <
> beowulf-boun...@beowulf.org on behalf of mikk...@mail.ru > wrote:

>> I worked always w/very small HPC clusters and built them manually (each 
>> server).
>> But what is reasonable to do for clusters containing some tens or hundred of
>> nodes ?
>> Of course w/modern Xeon (or Xeon Phi KNL) and IB EDR, during the next year 
>> for
>> example.
>> There are some automatic systems like OSCAR or even ROCKS.

>> But it looks that ROCKS don't support modern interconnects, and there may be
>> problems
>> w/OSCAR versions for support of systemd-based distributives like CentOS 7. 
>> For
>> next year -
>> is it reasonable to wait new OSCAR version or something else ?

>> Mikhail Kuzminsky,
>> Zelinsky Institute of Organic Chemistry RAS,
>> Moscow

> -- The Wellcome Trust Sanger Institute is operated by Genome Research 
> Limited, a
> charity registered in England with number 1021457 and a company registered in
> England with number 2742969, whose registered office is 215 Euston Road,
> London, NW1 2BE.
> _______________________________________________
> Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf

> --
> ___________________________________________

> _______________________________________________
> Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to