On Thursday, 3 May 2018 11:53:14 PM AEST John Hearns via Beowulf wrote: > The best successes I have seen on clusters is where the heavy parallel > applications get exclusive compute nodes. Cleaner, you get all the memory > and storage bandwidth and easy to clean up. Hell, reboot the things after > each job. You got an exclusive node.
You are describing the BlueGene/Q philosophy there John. :-) This idea tends to break when you throw GPUs in to the mix as there (hopefully) you only need a couple of cores on the node to shovel data around and the GPU does the gruntwork. That means you'll generally have cores left over that could be doing something useful. On the cluster I'm currently involved with we've got 36 cores per node and a pair of P100 GPUs. We have 2 Slurm partitions per node, one for non-GPU jobs that can only use up to 32 cores per node and another for GPU jobs that has no restriction. This means we always keep at least 4 cores per node free for GPU jobs. All the best, Chris -- Chris Samuel : http://www.csamuel.org/ : Melbourne, VIC _______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf