Oh, sure, sorry. We are using slurm 18.08.8, with a backfill scheduler. The jobs are being assigned to the same partition, which limits gpus and cpus to 1 via QOS. Here some of the main flags:
SallocDefaultCommand="srun -n1 -N1 --mem-per-cpu=0 --gres=gpu:0 --pty --preserve-env --mpi=none $SHELL" TaskPlugin=task/affinity,task/cgroup TaskPluginParam=Sched MinJobAge=300 FastSchedule=1 SchedulerType=sched/backfill SelectType=select/cons_res SelectTypeParameters=CR_CPU_Memory PreemptType=preempt/qos PreemptMode=requeue PriorityType=priority/multifactor PriorityFlags=FAIR_TREE PriorityFavorSmall=YES FairShareDampeningFactor=5 PriorityWeightAge=1000 PriorityWeightFairshare=5000 PriorityWeightJobSize=1000 PriorityWeightPartition=1000 PriorityWeightQOS=5000 PriorityWeightTRES=gres/gpu=1000 AccountingStorageEnforce=limits,qos,nosteps AccountingStorageTRES=gres/gpu AccountingStorageHost=localhost AccountingStorageType=accounting_storage/slurmdbd JobCompType=jobcomp/none JobAcctGatherFrequency=30 JobAcctGatherType=jobacct_gather/cgroup Any ideas? Cheers, El mié., 21 oct. 2020 a las 15:17, Riebs, Andy (<andy.ri...@hpe.com>) escribió: > Also, of course, any of the information that you can provide about how the > system is configured: scheduler choices, QOS options, and the like, would > also help in answering your question. > > > > *From:* slurm-users [mailto:slurm-users-boun...@lists.schedmd.com] *On > Behalf Of *Riebs, Andy > *Sent:* Wednesday, October 21, 2020 9:02 AM > *To:* Slurm User Community List <slurm-users@lists.schedmd.com> > *Subject:* Re: [slurm-users] Array jobs vs Fairshare > > > > Stephan (et al.), > > > > There are probably 6 versions of Slurm in common use today, across > multiple versions each of Debian/Ubuntu, SuSE/SLES, and > RedHat/CentOS/Fedora. You are more likely to get a good answer if you offer > some hints about what you are running! > > > > Regards, > > Andy > > > > *From:* slurm-users [mailto:slurm-users-boun...@lists.schedmd.com > <slurm-users-boun...@lists.schedmd.com>] *On Behalf Of *Stephan Schott > *Sent:* Wednesday, October 21, 2020 8:37 AM > *To:* Slurm User Community List <slurm-users@lists.schedmd.com> > *Subject:* [slurm-users] Array jobs vs Fairshare > > > > Hi everyone, > > I am having doubts regarding array jobs. To me it seems that the > JobArrayTaskLimit has precedence over the Fairshare, as users with a way > lower priority seem to get constant allocations for their array jobs, > compared to users with "normal" jobs. Can someone confirm this? > > Cheers, > > > -- > > Stephan Schott Verdugo > > Biochemist > > > Heinrich-Heine-Universitaet Duesseldorf > Institut fuer Pharm. und Med. Chemie > Universitaetsstr. 1 > 40225 Duesseldorf > Germany > -- Stephan Schott Verdugo Biochemist Heinrich-Heine-Universitaet Duesseldorf Institut fuer Pharm. und Med. Chemie Universitaetsstr. 1 40225 Duesseldorf Germany