Some examples are here:
https://wiki.fysik.dtu.dk/niflheim/Slurm_accounting#quality-of-service-qos
/Ole
On 19-12-2019 19:30, Prentice Bisbal wrote:
On 12/19/19 10:44 AM, Ransom, Geoffrey M. wrote:
The simplest is probably to just have a separate partition that will
only allow job times of 1
On 12/19/19 10:44 AM, Ransom, Geoffrey M. wrote:
The simplest is probably to just have a separate partition that will
only allow job times of 1 hour or less.
This is how our Univa queues used to work, by overlapping the same
hardware. Univa shows available “slots” to the users and we had a
So I’ve found some more info on this. It seems like the primary controller is
writing “ none” as the AllocNodes value in the partition state file when it
shuts down. It does this even with the backup out of the picture, and it still
happens even when I switched the primary and backup controlle
My current batch queues have a 30-day limit, and I’ll likely be reducing that
to maybe 7 days for most users in the near future, as it will make priority and
fairshare mechanisms more responsive (even if a high-priority job gets bumped
to the top of the queue, it may still have to wait a few day
The simplest is probably to just have a separate partition that will
only allow job times of 1 hour or less.
This is how our Univa queues used to work, by overlapping the same hardware.
Univa shows available "slots" to the users and we had a lot of confused users
complaining about al
Hello,
Don't use quotation marks with your command.
For example, you can execute:
srun hostname -f
but it won't work when you type:
srun "hostname -f"
That's because it would try to execute "hostname -f" executalbe, not
only "hostname" with "-f" as an argument.
Best regards,
Jacek
W dni