We run at about 50 PF and 1.5k nodes with about 100,000 jobs per day and we use 25.05.4, though we tend to upgrade to latest available so we will be upgrading to 25.11.* soon (when the .1 release comes out). If you are interested I'm happy to share our slurm.conf.

At least from my experience the latest releases have been stable, though you want to avoid .0 releases unless you want to be bleeding edge or need a feature. Most of the kinks are worked out by .1, definitely by .2 of any major release. There still may be weird edge cases but in general it is stable.

-Paul Edmon-

On 11/16/25 10:33 AM, John Hearns via slurm-users wrote:
I would take a step back and ask how you intend to install and manage this cluster.

CPU only or GPUs ?
OS ?
Interconnect fabric?
Storage ?

Power per rack? Cooling?
Monitoring?

On Sun, Nov 16, 2025, 2:39 PM KK via slurm-users <[email protected]> wrote:

    We are currently planning to deploy a new HPC system with a total
    compute capacity exceeding 100 PF. As part of our preparation, we
    would like to understand which Slurm versions are considered
    stable and widely used at this scale.

    Could you please share your recommendations or experience regarding:

    1. Which Slurm version is currently running reliably on very
    large-scale clusters (>100 PF or >10k nodes)?

    2. Whether there are any versions we should avoid due to known
    issues at large scale.

    3. Any best practices or configuration considerations for Slurm
    deployments of this size.


-- slurm-users mailing list -- [email protected]
    To unsubscribe send an email to [email protected]

-- 
slurm-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to