Re: [slurm-users] Not able to allocate all 24 ntasks-per-node; slurm.conf appears correct

2019-03-28 Thread Chris Samuel
Hi Anne, On 28/3/19 1:33 pm, Anne M. Hammond wrote: So one has to be explicit about ntasks-per-core on the command line or in the sbatch file, even though ThreadsPerCore is in slurm.conf. If I remember correctly by default Slurm will only launch one task per physical core, hence why you need

Re: [slurm-users] Not able to allocate all 24 ntasks-per-node; slurm.conf appears correct

2019-03-28 Thread Uwe Sauter
You probably shouldn't use "CPUs=24" and "Sockets=2 CoresPerSocket=6 ThreadsPerCore=2" at the same time. CPUs=24 (or more exactly Tasks=24, though this setting does not exist) follows directly from the other settings. From slurm.conf man page: CPUs Number of logical processors on the node

Re: [slurm-users] Not able to allocate all 24 ntasks-per-node; slurm.conf appears correct

2019-03-28 Thread Anne M. Hammond
Thanks Jeffrey and Chris. This does not work: srun -N 1 -n 24 --pty bash -i But this works: srun -N 1 -n 24 --ntasks-per-core=2 --pty bash -i So one has to be explicit about ntasks-per-core on the command line or in the sbatch file, even though ThreadsPerCore is in slurm.conf. > On Mar 2

Re: [slurm-users] Slurm doesn't call mpiexec or mpirun when run through a GUI app

2019-03-28 Thread Reuti
Hi, > Am 22.03.2019 um 16:20 schrieb Prentice Bisbal : > > On 3/21/19 6:56 PM, Reuti wrote: >> Am 21.03.2019 um 23:43 schrieb Prentice Bisbal: >> >>> Slurm-users, >>> >>> My users here have developed a GUI application which serves as a GUI >>> interface to various physics codes they use. From

Re: [slurm-users] Multinode MPI job

2019-03-28 Thread Mahmood Naderan
I test with env strace srun --pack-group=0 --ntasks=2 : --pack-group=1 --ntasks=4 pw.x -i mos2.rlx.in in the slurm script and everything is fine now!! This is going to be a nasty bug to find... Regards, Mahmood On Thu, Mar 28, 2019 at 9:18 PM Mahmood Naderan wrote: > Yes that works. > > $

Re: [slurm-users] Multinode MPI job

2019-03-28 Thread Mahmood Naderan
Yes that works. $ grep "Parallel version" big-mem Parallel version (MPI), running on 1 processors Parallel version (MPI), running on 1 processors Parallel version (MPI), running on 1 processors Parallel version (MPI), running on 1 processors $ squeue

Re: [slurm-users] Multinode MPI job

2019-03-28 Thread Frava
Well, does it also crash when you run it with two nodes in a normal way (not using heterogeneous jobs) ? #!/bin/bash #SBATCH --job-name=myQE_2Nx2MPI #SBATCH --output=big-mem #SBATCH --nodes=2 #SBATCH --ntasks-per-node=2 #SBATCH --mem-per-cpu=16g #SBATCH --partition=QUARTZ #SBATCH --account=z5 # sr

Re: [slurm-users] Multinode MPI job

2019-03-28 Thread Mahmood Naderan
BTW, when I manually run on a node, e.g. compute-0-2, I get this output ]$ mpirun -np 4 pw.x -i mos2.rlx.in Program PWSCF v.6.2 starts on 28Mar2019 at 11:40:36 This program is part of the open-source Quantum ESPRESSO suite for quantum simulation of materials; please cite

Re: [slurm-users] Multinode MPI job

2019-03-28 Thread Mahmood Naderan
The run is not consistent. I have manually test "mpirun -np 4 pw.x -i mos2.rlx.in" on compute-0-2 and rocks7 nodes and it is fine. However, with the script "srun --pack-group=0 --ntasks=2 : --pack-group=1 --ntasks=4 pw.x -i mos2.rlx.in" I see some errors in the output file which results in abortion

Re: [slurm-users] number of nodes varies for no reason?

2019-03-28 Thread Noam Bernstein
> On Mar 27, 2019, at 9:32 PM, Chris Samuel wrote: > > On 27/3/19 2:43 pm, Noam Bernstein wrote: > >> Hi fellow slurm users - I’ve been using slurm happily for a few months, but >> now I feel like it’s gone crazy, and I’m wondering if anyone can explain >> what’s going on. I have a trivial ba

Re: [slurm-users] spart: A user-oriented partition info command for slurm

2019-03-28 Thread Jeffrey Frey
Just FYI, there's one minor issue with spart: for pending jobs, the "partition" parameter can be a comma-separated list of partitions and not just a single partition name If the job can use more than one partition, specify their names in a comma separate list and the one offering earliest

Re: [slurm-users] Multinode MPI job

2019-03-28 Thread Frava
I didn't receive the last mail from Mahmood but Marcus is right, Mahmood's heterogeneous job submission seems to be working now. Well, separating each pack in the srun command and asking for the correct number of tasks to be launched for each pack is the way I figured the heterogeneous jobs worked

Re: [slurm-users] Multinode MPI job

2019-03-28 Thread Marcus Wagner
Hi Mahmood, On 3/28/19 7:33 AM, Mahmood Naderan wrote: >srun --pack-group=0 --ntasks=2 : --pack-group=1 --ntasks=4 pw.x -i mos2.rlx.in Still only one node is running the processes no, the processes are running as had been asked for. $ squeue JOBID PARTITION