[slurm-users] Re: Max TRES per user and node

2024-09-25 Thread Guillaume COCHARD via slurm-users
- Mail original - De: "Paul Raines" À: "Guillaume COCHARD" Cc: "Slurm User Community List" Envoyé: Mercredi 25 Septembre 2024 15:29:28 Objet: Re: [slurm-users] Re: Max TRES per user and node I am pretty sure there is no way to do exactly a per user per node limit

[slurm-users] Re: Max TRES per user and node

2024-09-25 Thread Paul Raines via slurm-users
-node limiting. Rob From: Carsten Beyer via slurm-users Sent: Wednesday, September 25, 2024 7:27 AM To: Guillaume COCHARD Cc: Slurm User Community List Subject: [slurm-users] Re: Max TRES per user and node Hi Guillaume, as Rob it already mentioned, this could

[slurm-users] Re: Max TRES per user and node

2024-09-25 Thread Groner, Rob via slurm-users
M To: Guillaume COCHARD Cc: Slurm User Community List Subject: [slurm-users] Re: Max TRES per user and node Hi Guillaume, as Rob it already mentioned, this could maybe a way for you (partition just created temporarily online for testing). You could also add your MaxTRES=node=1 for more rest

[slurm-users] Re: Max TRES per user and node

2024-09-25 Thread Carsten Beyer via slurm-users
Hi Guillaume, as Rob it already mentioned, this could maybe a way for you (partition just created temporarily online for testing). You could also add your MaxTRES=node=1 for more restrictions. We do something similar with QOS to restrict the number of CPU's for user in certain partitions.

[slurm-users] Re: Max TRES per user and node

2024-09-24 Thread Groner, Rob via slurm-users
Ok, that example helped. Max of 200G on a single node, per user (not job). No limits on how many jobs and nodes they can use...just a limit of 200G per node per user. And in that case, it's out of my realm of experience. 🙂 I'm relatively confident there IS a way...but I don't know it offha

[slurm-users] Re: Max TRES per user and node

2024-09-24 Thread Guillaume COCHARD via slurm-users
> "So if they submit a 2 nd job, that job can start but will have to go onto > another node, and will again be restricted to 200G? So they can start as many > jobs as there are nodes, and each job will be restricted to using 1 node and > 200G of memory?" Yes that's it. We already have MaxNodes

[slurm-users] Re: Max TRES per user and node

2024-09-24 Thread Groner, Rob via slurm-users
Ah, sorry, I didn't catch that from your first post (though you did say it). So, you are trying to limit the user to no more than 200G of memory on a single node? So if they submit a 2nd job, that job can start but will have to go onto another node, and will again be restricted to 200G? So the

[slurm-users] Re: Max TRES per user and node

2024-09-24 Thread Guillaume COCHARD via slurm-users
Thank you for your answer. To test it I tried: sacctmgr update qos normal set maxtresperuser=cpu=2 # Then in slurm.conf PartitionName=test […] qos=normal But then if I submit several 1-cpu jobs only two start and the others stay pending, even though I have several nodes available. So it see

[slurm-users] Re: Max TRES per user and node

2024-09-24 Thread Groner, Rob via slurm-users
You have the right idea. On that same page, you'll find MaxTRESPerUser, as a QOS parameter. You can create a QOS with the restrictions you'd like, and then in the partition definition, you give it that QOS. The QOS will then apply its restrictions to any jobs that use that partition. Rob