Hi all, I have a single server with 64 cores. The first 4 cores are reserved for slurm daemons (CPUSpecList=0-3) and for ssh connections. So slurm have 60 cores to assign jobs to.
I do not know why but at a certain time, the CPUAlloc value becomes incorrect and wrongly calculated. For example, right now, there are 68 cores assigned by slurm : squeue -o "%.18i %.8j %.2t %C" JOBID NAME ST CPUS 30975 map2R R 4 30976 map2R R 4 31016 bwa_mem_ R 8 31091 wgstest R 10 31225 wgstest R 10 31226 wgstest R 10 31227 wgstest R 10 31232 complete R 1 31265 repeatMM R 10 31295 LR2016Se R 1 But slurm thinks there are only 57: katak:~ # scontrol scontrol: show node NodeName=katak Arch=x86_64 CoresPerSocket=8 CPUAlloc=57 CPUErr=0 CPUTot=64 CPULoad=62.18 If I start a 10 cores task, it will start and the CPUAlloc value will be increased only by 1 (to 58). I think (but not sure) this wrong behavior starts when users forget to submit their jobs in slurm (so these jobs starts in the 4 cores reserved for slurm daemons). It makes slurm and ssh connections more unresponsive until I kill those jobs. Any help or advice appreciated, Thank you! --- Stéphane Larose Analyste de l'informatique Institut de Biologie Intégrative et des Systèmes (IBIS) Pavillon Charles-Eugène-Marchand Université Laval