[slurm-users] Re: Slurm not running on a warewulf node

2024-12-04 Thread Hermann Schwärzler via slurm-users
Hi Steven, yes, you have the syntax a bit wrong. If you consult the documentation (or the man-page) of slurm.conf you find this in the "NODE CONFIGURATION" section (in the paragraph about "NodeName"): Note that if the short form of the hostname is not used, it may prevent use of hostlist ex

[slurm-users] Re: Annoying canonical question about converting SLURM_JOB_NODELIST to a host list for mpirun

2024-08-09 Thread Hermann Schwärzler via slurm-users
Hi Paul, On 8/9/24 18:45, Paul Edmon via slurm-users wrote: As I recall I think OpenMPI needs a list that has an entry on each line, rather than one seperated by a space. See: [root@holy7c26401 ~]# echo $SLURM_JOB_NODELIST holy7c[26401-26405] [root@holy7c26401 ~]# scontrol show hostnames $SLUR

[slurm-users] Re: How to exclude master from computing? Set to DRAINED?

2024-06-24 Thread Hermann Schwärzler via slurm-users
Dear Xaver, we have a similar setup and yes, we have set the node to "state=DRAIN". Slurm keeps it this way until you manually change it to e.g. "state=RESUME". Regards, Hermann On 6/24/24 13:54, Xaver Stiensmeier via slurm-users wrote: Dear Slurm users, in our project we exclude the master f

[slurm-users] Re: sbatch and --nodes

2024-05-31 Thread Hermann Schwärzler via slurm-users
Hi Michael, if you submit a job-array, all resources related options (number of nodes, tasks, cpus per task, memory, time, ...) are meant *per array-task*. So in your case you start 100 array-tasks (you could also call them "sub-jobs") that *each* (not your whole job) is limited to one node, on

[slurm-users] Re: sbatch problem

2024-05-29 Thread Hermann Schwärzler via slurm-users
bde063af5c8e GPU-bf12181a-e615-dcd4-5da2-9a518ae1af5d GPU-dfec21c4-e30d-5a36-599d-eef2fd354809 GPU-15a11fe2-33f2-cd65-09f0-9897ba057a0c GPU-2d971e69-8147-8221-a055-e26573950f91 GPU-22ee3c89-fed1-891f-96bb-6bbf27a2cc4b Job finished at: Tue May 28 13:03:20 EEST 2024 ...I'm not interesting on th

[slurm-users] Re: sbatch problem

2024-05-28 Thread Hermann Schwärzler via slurm-users
18ae1af5d GPU-dfec21c4-e30d-5a36-599d-eef2fd354809 GPU-15a11fe2-33f2-cd65-09f0-9897ba057a0c GPU-2d971e69-8147-8221-a055-e26573950f91 GPU-22ee3c89-fed1-891f-96bb-6bbf27a2cc4b Job finished at: Tue May 28 13:03:20 EEST 2024 ...I'm not interesting on the output of the other 'echo' co

[slurm-users] Re: sbatch problem

2024-05-28 Thread Hermann Schwärzler via slurm-users
Hi Mihai, this is a problem that is not Slurm related. It's rather about: "when does command substitution happen?" When you write srun echo Running on host: $(hostname) $(hostname) is replaced by the output of the hostname-command *before* the line is "submitted" to srun. Which means that s

[slurm-users] Re: Performance Discrepancy between Slurm and Direct mpirun for VASP Jobs.

2024-05-27 Thread Hermann Schwärzler via slurm-users
Hi everbody, On 5/26/24 08:40, Ole Holm Nielsen via slurm-users wrote: [...] Whether or not to enable Hyper-Threading (HT) on your compute nodes depends entirely on the properties of applications that you wish to run on the nodes.  Some applications are faster without HT, others are faster wit

[slurm-users] Re: Performance Discrepancy between Slurm and Direct mpirun for VASP Jobs.

2024-05-24 Thread Hermann Schwärzler via slurm-users
Hi Zhao, my guess is that in your faster case you are using hyperthreading whereas in the Slurm case you don't. Can you check what performance you get when you add #SBATCH --hint=multithread to you slurm script? Another difference between the two might be a) the communication channel/interf

[slurm-users] Re: srun weirdness

2024-05-15 Thread Hermann Schwärzler via slurm-users
Hi Dj, could be a memory-limits related problem. What is the output of ulimit -l -m -v -s in both interactive job-shells? You are using cgroups-v1 now, right? In that case what is the respective content of /sys/fs/cgroup/memory/slurm_*/uid_$(id -u)/job_*/memory.limit_in_bytes in both shell

[slurm-users] Re: sbatch and cgroup v2

2024-02-28 Thread Hermann Schwärzler via slurm-users
Hi Dietmar, what do you find in the output-file of this job sbatch --time 5 --cpus-per-task=1 --wrap 'grep Cpus /proc/$$/status' On our 64 cores machines with enabled hyperthreading I see e.g. Cpus_allowed: 0400,,0400, Cpus_allowed_list: 58,122 Greetings Hermann