- Mail original -
De: "Paul Raines"
À: "Guillaume COCHARD"
Cc: "Slurm User Community List"
Envoyé: Mercredi 25 Septembre 2024 15:29:28
Objet: Re: [slurm-users] Re: Max TRES per user and node
I am pretty sure there is no way to do exactly a per user per node limit
-node limiting.
Rob
From: Carsten Beyer via slurm-users
Sent: Wednesday, September 25, 2024 7:27 AM
To: Guillaume COCHARD
Cc: Slurm User Community List
Subject: [slurm-users] Re: Max TRES per user and node
Hi Guillaume,
as Rob it already mentioned, this could
M
To: Guillaume COCHARD
Cc: Slurm User Community List
Subject: [slurm-users] Re: Max TRES per user and node
Hi Guillaume,
as Rob it already mentioned, this could maybe a way for you (partition just
created temporarily online for testing). You could also add your MaxTRES=node=1
for more rest
Hi Guillaume,
as Rob it already mentioned, this could maybe a way for you (partition
just created temporarily online for testing). You could also add your
MaxTRES=node=1 for more restrictions. We do something similar with QOS
to restrict the number of CPU's for user in certain partitions.
Ok, that example helped. Max of 200G on a single node, per user (not job). No
limits on how many jobs and nodes they can use...just a limit of 200G per node
per user.
And in that case, it's out of my realm of experience. 🙂 I'm relatively
confident there IS a way...but I don't know it offha
> "So if they submit a 2 nd job, that job can start but will have to go onto
> another node, and will again be restricted to 200G? So they can start as many
> jobs as there are nodes, and each job will be restricted to using 1 node and
> 200G of memory?"
Yes that's it. We already have MaxNodes
Ah, sorry, I didn't catch that from your first post (though you did say it).
So, you are trying to limit the user to no more than 200G of memory on a single
node? So if they submit a 2nd job, that job can start but will have to go onto
another node, and will again be restricted to 200G? So the
Thank you for your answer.
To test it I tried:
sacctmgr update qos normal set maxtresperuser=cpu=2
# Then in slurm.conf
PartitionName=test […] qos=normal
But then if I submit several 1-cpu jobs only two start and the others stay
pending, even though I have several nodes available. So it see
You have the right idea.
On that same page, you'll find MaxTRESPerUser, as a QOS parameter.
You can create a QOS with the restrictions you'd like, and then in the
partition definition, you give it that QOS. The QOS will then apply its
restrictions to any jobs that use that partition.
Rob