On Saturday, 14 April 2018 12:00:59 AM AEST Mahmood Naderan wrote:
> What are those? What should I mention in the format option to see the
> proper information about three lines of a jobs?
Yes. Even just running "sacct" on its own will tell you that information, the
reason it doesn't at the mom
On Saturday, 14 April 2018 1:33:13 AM AEST Mahmood Naderan wrote:
> I tried with one of the NAS benchmarks (BT) with 121 threads since the
> number of cores should be square.
That's an IO benchmark, not going to help you for this. You need something
that is compute bound & comms intensive to se
I respect this is a technical list and that SchedMD is running it so I will
say my bit once and keep it short but I think it's important to get these
ideas out there. These thoughts are mine and do not constitute any official
statement on the part of my employer.
My lab management probably likes u
Hi Philippe,
The built-in --x11 support still has some rough edges to it. Due to a
timeout issue with getting xauth set up, I've been patching SLURM with
these minor changes:
--- src/common/run_command.c_orig 2018-02-06 15:03:10.0 -0800
+++ src/common/run_command.c 2018-02-06 15:26:23.525
The output is certainly not enough to judge, but my first guess would be
that your MPI (what is it btw?) is not support PMI that is enabled in Slurm.
Note also, that Slurm now supports 3 ways of doing PMI and from the info
that you have provided it is not clear which one you are using.
To judge wit
Hi Chris,
I have been confused with the cpu runtime values in the sacct. For a
multinode mpi job, I see these values
[mahmood@rocks7 ~]$ sacct --format=jobid,user,cputime,elapsed,totalcpu,ncpus
JobID UserCPUTimeElapsed TotalCPU NCPUS
- -- ---
On 04/11/2018 02:35 PM, Sean Caron wrote:
As a protest to asking questions on this list and getting solicitations
for pay-for support, let me give you some advice for free :)
Now, now. Paid support is how they keep the project going. You like
using Slurm, right?
I tried with one of the NAS benchmarks (BT) with 121 threads since the
number of cores should be square. With srun, I get
WARNING: compiled for 121 processes
Number of active processes: 1
0 1 408 408 408
Problem size too big for compiled arra
On 13/4/18 7:19 pm, Mahmood Naderan wrote:
I see some old posts on the web about performance comparison of srun
vs. mpirun. Is that still an issue?
Just running an MPI hello world program is not going to test that.
You need to run an actual application that is doing a lot of
computation and c
I’m asking in the hopes that others will chime in (I’m curious why this is
happening)
Could you share your related slurm.conf cgroup options
cgroup.conf
cgroup_allowed_devices_file.conf
TaskPlugin
ProctrackType
JobAcctGatherType
-Kevin
PS Looking for similar style jobs, We have >1 day gpu use
On Fri, 13 Apr 2018 13:49:56 +0430
Mahmood Naderan wrote:
> Hi,
> I see some old posts on the web about performance comparison of srun
> vs. mpirun. Is that still an issue? Both the following scripts works
> for test programs and surely the performance concerns is not visible
> here.
...
> #SBAT
Excuse me, I see three lines also!
[root@rocks7 ~]# sacct --format=user,cputime,elapsed
UserCPUTimeElapsed
- -- --
mahmood 02:44:00 00:10:15
02:44:00 00:10:15
mahmood 19:25:52 00:18:13
09:42:56 00:18:13
00:18
Hi,
I see some old posts on the web about performance comparison of srun
vs. mpirun. Is that still an issue? Both the following scripts works
for test programs and surely the performance concerns is not visible
here.
#!/bin/bash
#SBATCH --job-name=hello_mpi
#SBATCH --output=hellompi.log
#SBATCH --
13 matches
Mail list logo