We have had a similar problem, even with different partitions for CPU
and GPU nodes, people still submitted jobs to the GPU nodes, and we
suspected running CPU type jobs. Doesn't help to look for the missing
--gres=gpu:x because a user can ask for GPUs and simply not use them. We
thought of get
Ward Poelmans wrote:
On 6/07/2021 14:59, Emre Brookes wrote:
I'm using slurm 20.02.7 & have the same issue (except I am running batch jobs).
Does MinJobAge work to keep completed jobs around for the specified duration in
squeue output?
It does for me if I do 'squeue -t all'. This is slurm 20.1
On 6/07/2021 14:59, Emre Brookes wrote:
> I'm using slurm 20.02.7 & have the same issue (except I am running batch
> jobs).
> Does MinJobAge work to keep completed jobs around for the specified duration
> in squeue output?
It does for me if I do 'squeue -t all'. This is slurm 20.11.7.
Ward
The documentation indicates that's what should happen with MinJobAge:
*MinJobAge*
The minimum age of a completed job before its record is purged from
Slurm's active database. Set the values of *MaxJobCount* and to
ensure the slurmctld daemon does not exhaust its memory or other
resour
Brian Andrus
Nov 23, 2020, 1:55:54 PM
to slurm...@lists.schedmd.com
All,
I always thought that MinJobAge affected how long a job will show up
when doing 'squeue'
That does not seem to be the case for me.
I have MinJobAge=900, but if I do 'squeue --me' as soon as I finish an
interactive