[slurm-users] Re: Suggestions for Partition/QoS configuration

2024-04-04 Thread thomas.hartmann--- via slurm-users
Hi, I'm currently testing an approach similar to the example by Loris. Why consider preemption? Because, in the original example, if the cluster is saturated by long running jobs (like 2 weeks), there should be the possibility to run short jobs right away. Best, Thomas -- slurm-users mailing

[slurm-users] Suggestions for Partition/QoS configuration

2024-04-04 Thread thomas.hartmann--- via slurm-users
Hi, we're testing possible slurm configurations on a test system right now. Eventually, it is going to serve ~1000 users. We're going to have some users who are going to run lots of short jobs (a couple of minutes to ~4h) and some users that run jobs that are going to run for days or weeks. I w

[slurm-users] Re: pty jobs are killed when another job on the same node terminates

2024-03-01 Thread thomas.hartmann--- via slurm-users
Hi, sorry, I had written an email but it apparently didn't go through Götz was right. slurm.epilog.clean was the problem. There was a bug in there... I fixed it and now it works. Best, Thomas -- slurm-users mailing list -- slurm-users@lists.schedmd.com To unsubscribe send an email to slurm

[slurm-users] How to get usage data for a QOS

2024-02-28 Thread thomas.hartmann--- via slurm-users
Hi, so, I figured out that I can give some users priority access for a specific amount of TRES by creating a qos with the GrpTRESMins property and the DenyOnLimit,NoDecay flags. This works nicely. However, I would like to know, how much of this has already been consumed and I have not yet found