On 09/28/2016 07:34 AM, Mikhail Kuzminsky wrote: > I worked always w/very small HPC clusters and built them manually > (each server).
Manual installs aren't too bad up to 4 nodes or so. > But what is reasonable to do for clusters containing some tens or > hundred of nodes ? We use cobbler for DHCP, bootp, DNS, and PXE boot. It's nice to have a single database for mac address, IP address, hostname, etc. We have a profile per OS. Then we use cobbler to netboot CentOS or Debian family OSs and part of the installation. The installation installs puppet which handles: * Which users can login to which hardware * Distribution of ssh keys * Installation of packages, services, monitoring, etc. * Tweaking initd/systemd scripts, pam, ulimit, etc. * Managing autofs and friends. > But it looks that ROCKS don't support modern interconnects, and there > may be problems > w/OSCAR versions for support of systemd-based distributives like CentOS > 7. For next year - > is it reasonable to wait new OSCAR version or something else ? The hard part of supporting an HPC side of things is the users/apps, I think of installation and configuration of hardware fairly minor. Personally I'd just pick an OS that best suites your user/application needs. PXE boot + cobbler with whatever linux OS is really not a big deal. With the above it's easy to write a small script to shutdown a node, turn on netboot, power on the node (assuming IPMI works), install on boot, reboot into OS, NFS mount, run slurmd daemon, and be back in production. _______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf