And I forgot to mention, things are running in a Qlustar cluster based on Ubuntu 18.04.4 LTS Bionic. š¬
El mié., 21 oct. 2020 a las 15:38, Stephan Schott (<schot...@hhu.de>) escribió: > Oh, sure, sorry. > We are using slurm 18.08.8, with a backfill scheduler. The jobs are being > assigned to the same partition, which limits gpus and cpus to 1 via QOS. > Here some of the main flags: > > SallocDefaultCommand="srun -n1 -N1 --mem-per-cpu=0 --gres=gpu:0 --pty > --preserve-env --mpi=none $SHELL" > TaskPlugin=task/affinity,task/cgroup > TaskPluginParam=Sched > MinJobAge=300 > FastSchedule=1 > SchedulerType=sched/backfill > SelectType=select/cons_res > SelectTypeParameters=CR_CPU_Memory > PreemptType=preempt/qos > PreemptMode=requeue > PriorityType=priority/multifactor > PriorityFlags=FAIR_TREE > PriorityFavorSmall=YES > FairShareDampeningFactor=5 > PriorityWeightAge=1000 > PriorityWeightFairshare=5000 > PriorityWeightJobSize=1000 > PriorityWeightPartition=1000 > PriorityWeightQOS=5000 > PriorityWeightTRES=gres/gpu=1000 > AccountingStorageEnforce=limits,qos,nosteps > AccountingStorageTRES=gres/gpu > AccountingStorageHost=localhost > AccountingStorageType=accounting_storage/slurmdbd > JobCompType=jobcomp/none > JobAcctGatherFrequency=30 > JobAcctGatherType=jobacct_gather/cgroup > > Any ideas? > > Cheers, > > El mié., 21 oct. 2020 a las 15:17, Riebs, Andy (<andy.ri...@hpe.com>) > escribió: > >> Also, of course, any of the information that you can provide about how >> the system is configured: scheduler choices, QOS options, and the like, >> would also help in answering your question. >> >> >> >> *From:* slurm-users [mailto:slurm-users-boun...@lists.schedmd.com] *On >> Behalf Of *Riebs, Andy >> *Sent:* Wednesday, October 21, 2020 9:02 AM >> *To:* Slurm User Community List <slurm-users@lists.schedmd.com> >> *Subject:* Re: [slurm-users] Array jobs vs Fairshare >> >> >> >> Stephan (et al.), >> >> >> >> There are probably 6 versions of Slurm in common use today, across >> multiple versions each of Debian/Ubuntu, SuSE/SLES, and >> RedHat/CentOS/Fedora. You are more likely to get a good answer if you offer >> some hints about what you are running! >> >> >> >> Regards, >> >> Andy >> >> >> >> *From:* slurm-users [mailto:slurm-users-boun...@lists.schedmd.com >> <slurm-users-boun...@lists.schedmd.com>] *On Behalf Of *Stephan Schott >> *Sent:* Wednesday, October 21, 2020 8:37 AM >> *To:* Slurm User Community List <slurm-users@lists.schedmd.com> >> *Subject:* [slurm-users] Array jobs vs Fairshare >> >> >> >> Hi everyone, >> >> I am having doubts regarding array jobs. To me it seems that the >> JobArrayTaskLimit has precedence over the Fairshare, as users with a way >> lower priority seem to get constant allocations for their array jobs, >> compared to users with "normal" jobs. Can someone confirm this? >> >> Cheers, >> >> >> -- >> >> Stephan Schott Verdugo >> >> Biochemist >> >> >> Heinrich-Heine-Universitaet Duesseldorf >> Institut fuer Pharm. und Med. Chemie >> Universitaetsstr. 1 >> 40225 Duesseldorf >> Germany >> > > > -- > Stephan Schott Verdugo > Biochemist > > Heinrich-Heine-Universitaet Duesseldorf > Institut fuer Pharm. und Med. Chemie > Universitaetsstr. 1 > 40225 Duesseldorf > Germany > -- Stephan Schott Verdugo Biochemist Heinrich-Heine-Universitaet Duesseldorf Institut fuer Pharm. und Med. Chemie Universitaetsstr. 1 40225 Duesseldorf Germany