Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread Arsene Marian Alain
Hello Brian,

Thanks for your answer. With the job_container/tmpfs plugin I don't really 
create the directory manually.

I just give my Basepath=/scratch (a local directory for each node that is 
already mounted with 1777 permissions) in job_container.conf. The plugin 
automatically generates for each job a directory with the "JOB_ID", for 
example: /scratch/1805

The only problem is that directory 1805 is generated with root owner and 
permissions 700. So the user who submitted the job cannot write/read inside 
directory 1805.

Is there a way for the owner of directory 1805 to be the user who submitted the 
job and not root?

De: slurm-users  En nombre de Brian 
Andrus
Enviado el: lunes, 20 de noviembre de 2023 23:29
Para: slurm-users@lists.schedmd.com
Asunto: Re: [slurm-users] slurm job_container/tmpfs

ATENCIÓN: Este correo electrónico se envió desde fuera de la UAH. No haga clic 
en enlaces ni abra archivos adjuntos a menos que reconozca al remitente y sepa 
que el contenido es seguro.

How do you 'manually create a directory'? That would be when the ownership of 
root would be occurring. After creating it, you can chown/chmod it as well.

Brian Andrus
On 11/18/2023 7:35 AM, Arsene Marian Alain wrote:
Dear slurm community,

I run slurm 21.08.1 under Rocky Linux 8.5 on my small HPC cluster and am trying 
to configure job_container/tmpfs to manage the temporary directories.

I have a shared nfs drive "/home" and a local "/scratch" (with permissions 
1777) on each node.

For each submitted job I manually create a directory with the "JOB_ID.$USER" in 
the local "/scratch" which is where all the temp files for the job will be 
generated. Now, I would like to do these automatically (especially to remove 
the directory when the job finishes or is canceled):

I added the following parameters in my /etc/slurm.conf:

JobContainerType=job_container/tmpfs
PrologFlags=contain

So, I have created the "job_container.conf" in the directory "/etc/slurm"
with the following configuration:

AutoBasePath=false
BasePath=/scratch

Then, I replicated the changes to all nodes and restarted the slurm daemons.

Finally, when I launch the job a directory with the "JOB_ID" is created in the 
local "/scratch" of the compute node. The only problem is that the owner of the 
directory is "root" and the user who submitted the job doesn’t have read and 
write permissions to that directory (other users do not either).

I would like that:

1) The name of the automatically created directory will be: "JOB_ID.$USER"
2) The owner of the directory will be the user who submitted the job, not 
"root".

Please, could someone help me?



Thanks a lot.

Best regards,

Alain


Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread Sean Mc Grath
Would a prolog script, https://slurm.schedmd.com/prolog_epilog.html, do what 
you need? Sorry if you have already considered that and I missed it.

---
Sean McGrath
Senior Systems Administrator, IT Services


From: slurm-users  on behalf of Arsene 
Marian Alain 
Sent: Tuesday 21 November 2023 09:58
To: Slurm User Community List 
Subject: Re: [slurm-users] slurm job_container/tmpfs


Hello Brian,



Thanks for your answer. With the job_container/tmpfs plugin I don't really 
create the directory manually.



I just give my Basepath=/scratch (a local directory for each node that is 
already mounted with 1777 permissions) in job_container.conf. The plugin 
automatically generates for each job a directory with the "JOB_ID", for 
example: /scratch/1805



The only problem is that directory 1805 is generated with root owner and 
permissions 700. So the user who submitted the job cannot write/read inside 
directory 1805.



Is there a way for the owner of directory 1805 to be the user who submitted the 
job and not root?



De: slurm-users  En nombre de Brian 
Andrus
Enviado el: lunes, 20 de noviembre de 2023 23:29
Para: slurm-users@lists.schedmd.com
Asunto: Re: [slurm-users] slurm job_container/tmpfs



ATENCIÓN: Este correo electrónico se envió desde fuera de la UAH. No haga clic 
en enlaces ni abra archivos adjuntos a menos que reconozca al remitente y sepa 
que el contenido es seguro.

How do you 'manually create a directory'? That would be when the ownership of 
root would be occurring. After creating it, you can chown/chmod it as well.

Brian Andrus

On 11/18/2023 7:35 AM, Arsene Marian Alain wrote:

Dear slurm community,



I run slurm 21.08.1 under Rocky Linux 8.5 on my small HPC cluster and am trying 
to configure job_container/tmpfs to manage the temporary directories.



I have a shared nfs drive "/home" and a local "/scratch" (with permissions 
1777) on each node.



For each submitted job I manually create a directory with the "JOB_ID.$USER" in 
the local "/scratch" which is where all the temp files for the job will be 
generated. Now, I would like to do these automatically (especially to remove 
the directory when the job finishes or is canceled):



I added the following parameters in my /etc/slurm.conf:



JobContainerType=job_container/tmpfs

PrologFlags=contain



So, I have created the "job_container.conf" in the directory "/etc/slurm"

with the following configuration:



AutoBasePath=false

BasePath=/scratch



Then, I replicated the changes to all nodes and restarted the slurm daemons.



Finally, when I launch the job a directory with the "JOB_ID" is created in the 
local "/scratch" of the compute node. The only problem is that the owner of the 
directory is "root" and the user who submitted the job doesn’t have read and 
write permissions to that directory (other users do not either).



I would like that:



1) The name of the automatically created directory will be: "JOB_ID.$USER"

2) The owner of the directory will be the user who submitted the job, not 
"root".



Please, could someone help me?







Thanks a lot.



Best regards,



Alain


Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread Ward Poelmans

Hi Arsene,

On 21/11/2023 10:58, Arsene Marian Alain wrote:


I just give my Basepath=/scratch (a local directory for each node that is already mounted 
with 1777 permissions) in job_container.conf. The plugin automatically generates for each 
job a directory with the "JOB_ID", for example: /scratch/1805

The only problem is that directory 1805 is generated with root owner and 
permissions 700. So the user who submitted the job cannot write/read inside 
directory 1805.


If I look on our system, there should be hidden directory under that directory 
which is owned by the correct user.

Our job_container.conf has:
Basepath=/local
Dirs=/var/tmp,/tmp,/dev/shm

which gives directories like:
/local/6000522/.6000522/_tmp/

The hidden one is owned by the user of the job.


Ward


smime.p7s
Description: S/MIME Cryptographic Signature


Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread Arsene Marian Alain
Hi Ward, 

You're right. 

[root@node01 scratch]# pwd
/scratch
[root@node01 scratch]# ll
total 0
drwx-- 3 root root 30 nov 21 13:41 1809
[root@node01 scratch]# ls -la 1809/
total 0
drwx--  3 root  root 30 nov 21 13:41 .
drwxrwxrwt. 3 root  root 18 nov 21 13:41 ..
drwx--  2  thais root  6 nov 21 13:41 .1809
-r--r--r--  1 root  root  0 nov 21 13:41 .ns

But how can user write or access the hidden directory .1809 if he doesn't have 
read/write permission on  main directory 1809?

Thanks.

-Mensaje original-
De: slurm-users  En nombre de Ward 
Poelmans
Enviado el: martes, 21 de noviembre de 2023 13:12
Para: slurm-users@lists.schedmd.com
Asunto: Re: [slurm-users] slurm job_container/tmpfs

Hi Arsene,

On 21/11/2023 10:58, Arsene Marian Alain wrote:

> I just give my Basepath=/scratch (a local directory for each node that is 
> already mounted with 1777 permissions) in job_container.conf. The plugin 
> automatically generates for each job a directory with the "JOB_ID", for 
> example: /scratch/1805
> 
> The only problem is that directory 1805 is generated with root owner and 
> permissions 700. So the user who submitted the job cannot write/read inside 
> directory 1805.

If I look on our system, there should be hidden directory under that directory 
which is owned by the correct user.

Our job_container.conf has:
Basepath=/local
Dirs=/var/tmp,/tmp,/dev/shm

which gives directories like:
/local/6000522/.6000522/_tmp/

The hidden one is owned by the user of the job.


Ward


Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread Arsene Marian Alain
Thanks Sean. I've tried using slurm prolog/epilog scripts but without any 
success. That's why I decided to look for other solutions and 
job_container/tmpfs plugin seemed like a good alternative.

De: slurm-users  En nombre de Sean Mc 
Grath
Enviado el: martes, 21 de noviembre de 2023 12:57
Para: Slurm User Community List 
Asunto: Re: [slurm-users] slurm job_container/tmpfs

ATENCIÓN: Este correo electrónico se envió desde fuera de la UAH. No haga clic 
en enlaces ni abra archivos adjuntos a menos que reconozca al remitente y sepa 
que el contenido es seguro.
Would a prolog script, https://slurm.schedmd.com/prolog_epilog.html, do what 
you need? Sorry if you have already considered that and I missed it.

---
Sean McGrath
Senior Systems Administrator, IT Services


From: slurm-users 
mailto:slurm-users-boun...@lists.schedmd.com>>
 on behalf of Arsene Marian Alain 
mailto:alain.ars...@uah.es>>
Sent: Tuesday 21 November 2023 09:58
To: Slurm User Community List 
mailto:slurm-users@lists.schedmd.com>>
Subject: Re: [slurm-users] slurm job_container/tmpfs


Hello Brian,



Thanks for your answer. With the job_container/tmpfs plugin I don't really 
create the directory manually.



I just give my Basepath=/scratch (a local directory for each node that is 
already mounted with 1777 permissions) in job_container.conf. The plugin 
automatically generates for each job a directory with the "JOB_ID", for 
example: /scratch/1805



The only problem is that directory 1805 is generated with root owner and 
permissions 700. So the user who submitted the job cannot write/read inside 
directory 1805.



Is there a way for the owner of directory 1805 to be the user who submitted the 
job and not root?



De: slurm-users 
mailto:slurm-users-boun...@lists.schedmd.com>>
 En nombre de Brian Andrus
Enviado el: lunes, 20 de noviembre de 2023 23:29
Para: slurm-users@lists.schedmd.com
Asunto: Re: [slurm-users] slurm job_container/tmpfs



ATENCIÓN: Este correo electrónico se envió desde fuera de la UAH. No haga clic 
en enlaces ni abra archivos adjuntos a menos que reconozca al remitente y sepa 
que el contenido es seguro.

How do you 'manually create a directory'? That would be when the ownership of 
root would be occurring. After creating it, you can chown/chmod it as well.

Brian Andrus

On 11/18/2023 7:35 AM, Arsene Marian Alain wrote:

Dear slurm community,



I run slurm 21.08.1 under Rocky Linux 8.5 on my small HPC cluster and am trying 
to configure job_container/tmpfs to manage the temporary directories.



I have a shared nfs drive "/home" and a local "/scratch" (with permissions 
1777) on each node.



For each submitted job I manually create a directory with the "JOB_ID.$USER" in 
the local "/scratch" which is where all the temp files for the job will be 
generated. Now, I would like to do these automatically (especially to remove 
the directory when the job finishes or is canceled):



I added the following parameters in my /etc/slurm.conf:



JobContainerType=job_container/tmpfs

PrologFlags=contain



So, I have created the "job_container.conf" in the directory "/etc/slurm"

with the following configuration:



AutoBasePath=false

BasePath=/scratch



Then, I replicated the changes to all nodes and restarted the slurm daemons.



Finally, when I launch the job a directory with the "JOB_ID" is created in the 
local "/scratch" of the compute node. The only problem is that the owner of the 
directory is "root" and the user who submitted the job doesn't have read and 
write permissions to that directory (other users do not either).



I would like that:



1) The name of the automatically created directory will be: "JOB_ID.$USER"

2) The owner of the directory will be the user who submitted the job, not 
"root".



Please, could someone help me?







Thanks a lot.



Best regards,



Alain


Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread René Sitt

Hello Alain,

as an alternative to job_container/tmpfs, you may also try your luck 
with the 'auto_tmpdir' SPANK plugin: 
https://github.com/University-of-Delaware-IT-RCI/auto_tmpdir


We've been using using that on our small HPC cluster (Slurm 22.05) and 
it does what it's supposed to. One thing one has to remember is that it 
requires a recompile after every Slurm update.


Kind regards,
René Sitt

Am 21.11.23 um 14:07 schrieb Arsene Marian Alain:


Thanks Sean. I’ve tried using slurm prolog/epilog scripts but without 
any success. That's why I decided to look for other solutions and 
job_container/tmpfs plugin seemed like a good alternative.


*De:* slurm-users  *En nombre 
de *Sean Mc Grath

*Enviado el:* martes, 21 de noviembre de 2023 12:57
*Para:* Slurm User Community List 
*Asunto:* Re: [slurm-users] slurm job_container/tmpfs

*ATENCIÓN*:Este correo electrónico se envió desde fuera de la UAH. No 
haga clic en enlaces ni abra archivos adjuntos a menos que reconozca 
al remitente y sepa que el contenido es seguro.


Would a prolog script, https://slurm.schedmd.com/prolog_epilog.html, 
do what you need? Sorry if you have already considered that and I 
missed it.


---

Sean McGrath

Senior Systems Administrator, IT Services



*From:*slurm-users  on behalf 
of Arsene Marian Alain 

*Sent:* Tuesday 21 November 2023 09:58
*To:* Slurm User Community List 
*Subject:* Re: [slurm-users] slurm job_container/tmpfs

Hello Brian,

Thanks for your answer. With the job_container/tmpfs plugin I don't 
really create the directory manually.


I just give my Basepath=/scratch (a local directory for each node that 
is already mounted with 1777 permissions) in job_container.conf. The 
plugin automatically generates for each job a directory with the 
"JOB_ID", for example: /scratch/1805


The only problem is that directory 1805 is generated with root owner 
and permissions 700. So the user who submitted the job cannot 
write/read inside directory 1805.


Is there a way for the owner of directory 1805 to be the user who 
submitted the job and not root?


*De:* slurm-users  *En nombre 
de *Brian Andrus

*Enviado el:* lunes, 20 de noviembre de 2023 23:29
*Para:* slurm-users@lists.schedmd.com
*Asunto:* Re: [slurm-users] slurm job_container/tmpfs

*ATENCIÓN*:Este correo electrónico se envió desde fuera de la UAH. No 
haga clic en enlaces ni abra archivos adjuntos a menos que reconozca 
al remitente y sepa que el contenido es seguro.


How do you 'manually create a directory'? That would be when the 
ownership of root would be occurring. After creating it, you can 
chown/chmod it as well.


Brian Andrus

On 11/18/2023 7:35 AM, Arsene Marian Alain wrote:

Dear slurm community,

I run slurm 21.08.1 under Rocky Linux 8.5 on my small HPC cluster
and am trying to configure job_container/tmpfs to manage the
temporary directories.

I have a shared nfs drive "/home" and a local "/scratch" (with
permissions 1777) on each node.

For each submitted job I manually create a directory with the
"JOB_ID.$USER" in the local "/scratch" which is where all the temp
files for the job will be generated. Now, I would like to do these
automatically (especially to remove the directory when the job
finishes or is canceled):

I added the following parameters in my /etc/slurm.conf:

JobContainerType=job_container/tmpfs

PrologFlags=contain

So, I have created the "job_container.conf" in the directory
"/etc/slurm"

with the following configuration:

AutoBasePath=false

BasePath=/scratch

Then, I replicated the changes to all nodes and restarted the
slurm daemons.

Finally, when I launch the job a directory with the "JOB_ID" is
created in the local "/scratch" of the compute node. The only
problem is that the owner of the directory is "root" and the user
who submitted the job doesn’t have read and write permissions to
that directory (other users do not either).

I would like that:

1) The name of the automatically created directory will be:
"JOB_ID.$USER"

2) The owner of the directory will be the user who submitted the
job, not "root".

Please, could someone help me?

Thanks a lot.

Best regards,

Alain


--
Dipl.-Chem. René Sitt
Hessisches Kompetenzzentrum für Hochleistungsrechnen
Philipps-Universität Marburg
Hans-Meerwein-Straße
35032 Marburg

Tel. +49 6421 28 23523
si...@hrz.uni-marburg.de
www.hkhlr.de


smime.p7s
Description: Kryptografische S/MIME-Signatur


Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread Lorenzo Bosio

Hello Alain,

maybe I'm missing the point, but from my understanding the 
job_container/tmpfs plugin uses the directory under BasePath to store 
its data, used to create the bind mounts for the users. The folder 
itself is not meant to be used by others.
The folders in the hidden directory with user privileges under your 
/scratch are the bind mounts. Those folders are specified in the Dirs 
parameter  
of job_container.conf. You may have more luck trying to use this 
parameter for your needs, perhaps? There is also a parameter to specify 
an "InitScript" which may be used to create folders dinamically.
One last thing, the those configuration has been added in one of the 
latest releases of Slurm, so they may not work with your version.


Best regards,
Lorenzo Bosio

On 21/11/23 14:07, Arsene Marian Alain wrote:


Thanks Sean. I’ve tried using slurm prolog/epilog scripts but without 
any success. That's why I decided to look for other solutions and 
job_container/tmpfs plugin seemed like a good alternative.


*De:* slurm-users  *En nombre 
de *Sean Mc Grath

*Enviado el:* martes, 21 de noviembre de 2023 12:57
*Para:* Slurm User Community List 
*Asunto:* Re: [slurm-users] slurm job_container/tmpfs

*ATENCIÓN*:Este correo electrónico se envió desde fuera de la UAH. No 
haga clic en enlaces ni abra archivos adjuntos a menos que reconozca 
al remitente y sepa que el contenido es seguro.


Would a prolog script, https://slurm.schedmd.com/prolog_epilog.html, 
do what you need? Sorry if you have already considered that and I 
missed it.


---

Sean McGrath

Senior Systems Administrator, IT Services



*From:*slurm-users  on behalf 
of Arsene Marian Alain 

*Sent:* Tuesday 21 November 2023 09:58
*To:* Slurm User Community List 
*Subject:* Re: [slurm-users] slurm job_container/tmpfs

Hello Brian,

Thanks for your answer. With the job_container/tmpfs plugin I don't 
really create the directory manually.


I just give my Basepath=/scratch (a local directory for each node that 
is already mounted with 1777 permissions) in job_container.conf. The 
plugin automatically generates for each job a directory with the 
"JOB_ID", for example: /scratch/1805


The only problem is that directory 1805 is generated with root owner 
and permissions 700. So the user who submitted the job cannot 
write/read inside directory 1805.


Is there a way for the owner of directory 1805 to be the user who 
submitted the job and not root?


*De:* slurm-users  *En nombre 
de *Brian Andrus

*Enviado el:* lunes, 20 de noviembre de 2023 23:29
*Para:* slurm-users@lists.schedmd.com
*Asunto:* Re: [slurm-users] slurm job_container/tmpfs

*ATENCIÓN*:Este correo electrónico se envió desde fuera de la UAH. No 
haga clic en enlaces ni abra archivos adjuntos a menos que reconozca 
al remitente y sepa que el contenido es seguro.


How do you 'manually create a directory'? That would be when the 
ownership of root would be occurring. After creating it, you can 
chown/chmod it as well.


Brian Andrus

On 11/18/2023 7:35 AM, Arsene Marian Alain wrote:

Dear slurm community,

I run slurm 21.08.1 under Rocky Linux 8.5 on my small HPC cluster
and am trying to configure job_container/tmpfs to manage the
temporary directories.

I have a shared nfs drive "/home" and a local "/scratch" (with
permissions 1777) on each node.

For each submitted job I manually create a directory with the
"JOB_ID.$USER" in the local "/scratch" which is where all the temp
files for the job will be generated. Now, I would like to do these
automatically (especially to remove the directory when the job
finishes or is canceled):

I added the following parameters in my /etc/slurm.conf:

JobContainerType=job_container/tmpfs

PrologFlags=contain

So, I have created the "job_container.conf" in the directory
"/etc/slurm"

with the following configuration:

AutoBasePath=false

BasePath=/scratch

Then, I replicated the changes to all nodes and restarted the
slurm daemons.

Finally, when I launch the job a directory with the "JOB_ID" is
created in the local "/scratch" of the compute node. The only
problem is that the owner of the directory is "root" and the user
who submitted the job doesn’t have read and write permissions to
that directory (other users do not either).

I would like that:

1) The name of the automatically created directory will be:
"JOB_ID.$USER"

2) The owner of the directory will be the user who submitted the
job, not "root".

Please, could someone help me?

Thanks a lot.

Best regards,

Alain


--
*/Dott. Mag. Lorenzo Bosio/*
Tecnico di Ricerca
Dipartimento di Informatica


Università degli Studi di Torino
Corso Svizzera, 185 - 10149 Torino
tel. +39 011 670 6836

Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread Roberto Monti
Hi,
From the perspective of the job, those directories are mapped to /tmp (and 
others, depending on your job_container.conf). There's no need for the user to 
be aware of the basepath that is specified in the conf file.

You can easily verify it is working by writing files to /tmp from a new slurm 
job, and running `find /scratch/` as root on the same node (while the 
job is still running).

Best,

--
Roberto P. Monti
DevOps Engineer I
roberto.mo...@jax.org

The Jackson Laboratory
United States | China | Japan
www.jax.org

-Original Message-
From: slurm-users  On Behalf Of Arsene 
Marian Alain
Sent: Tuesday, November 21, 2023 7:53 AM
To: Slurm User Community List 
Subject: [EXTERNAL]Re: [slurm-users] slurm job_container/tmpfs

Hi Ward, 

You're right. 

[root@node01 scratch]# pwd
/scratch
[root@node01 scratch]# ll
total 0
drwx-- 3 root root 30 nov 21 13:41 1809
[root@node01 scratch]# ls -la 1809/
total 0
drwx--  3 root  root 30 nov 21 13:41 .
drwxrwxrwt. 3 root  root 18 nov 21 13:41 ..
drwx--  2  thais root  6 nov 21 13:41 .1809
-r--r--r--  1 root  root  0 nov 21 13:41 .ns

But how can user write or access the hidden directory .1809 if he doesn't have 
read/write permission on  main directory 1809?

Thanks.

-Mensaje original-
De: slurm-users  En nombre de Ward 
Poelmans Enviado el: martes, 21 de noviembre de 2023 13:12
Para: slurm-users@lists.schedmd.com
Asunto: Re: [slurm-users] slurm job_container/tmpfs

Hi Arsene,

On 21/11/2023 10:58, Arsene Marian Alain wrote:

> I just give my Basepath=/scratch (a local directory for each node that 
> is already mounted with 1777 permissions) in job_container.conf. The 
> plugin automatically generates for each job a directory with the 
> "JOB_ID", for example: /scratch/1805
> 
> The only problem is that directory 1805 is generated with root owner and 
> permissions 700. So the user who submitted the job cannot write/read inside 
> directory 1805.

If I look on our system, there should be hidden directory under that directory 
which is owned by the correct user.

Our job_container.conf has:
Basepath=/local
Dirs=/var/tmp,/tmp,/dev/shm

which gives directories like:
/local/6000522/.6000522/_tmp/

The hidden one is owned by the user of the job.


Ward
---

The information in this email, including attachments, may be confidential and 
is intended solely for the addressee(s). If you believe you received this email 
by mistake, please notify the sender by return email as soon as possible.


Re: [slurm-users] slurm job_container/tmpfs

2023-11-21 Thread Ward Poelmans

Hi,

On 21/11/2023 13:52, Arsene Marian Alain wrote:



But how can user write or access the hidden directory .1809 if he doesn't have 
read/write permission on  main directory 1809?


Because it works as a namespace. On my side:

$ ls -alh /local/6000523/
total 0
drwx-- 3 root root 33 Nov 21 15:00 .
drwxrwxrwt 3 root root 21 Nov 21 15:00 ..
drwx-- 4 myuser   root 34 Nov 21 15:00 .6000523
-r--r--r-- 1 root root  0 Nov 21 15:00 .ns

While inside the job you just see this subdirectory.

Ward


smime.p7s
Description: S/MIME Cryptographic Signature


Re: [slurm-users] partition qos without managing users

2023-11-21 Thread egonle
ok, I understand synching of users to slurm database is a task which it 
not built-in, but could be added outside of slurm :-)


With regards to the QoS or Partition QoS setting I've tried several 
settings and configurations however it was not possible at all to 
configure a QoS on partition level only without not adding specific 
users to the slurm database.
Either I don't understand the docs properly or there is no configuration 
option to limit jobs with e.g. cpu=4 globally on a partition.


Could anybody share a configuration which set partition QoS (e.g. cpu=8) 
without managing users or a configuration to silently change the job QoS 
using job_submit.lua again without maintaining users within slurm 
database?



Thanks


Date: Mon, 20 Nov 2023 14:37:11 -0800
From: Brian Andrus 
To: slurm-users@lists.schedmd.com
Subject: Re: [slurm-users] partition qos without managing users
Message-ID: <2f421687-40aa-4e35-bf9d-3f31984ad...@gmail.com>
Content-Type: text/plain; charset=UTF-8; format=flowed

You would have to do such syncing with your own scripts. There is no 
way

slurm would be able to tell which users should have access and what
access without the slurmdb and such info is not contained in AD.

At our site, we iterate through the group(s) that are slurm user groups
and add the users if they do not exist. We also delete users when they
are removed from AD. This does have the effect of losing job info
produced by said users, but since we export that into a larger historic
repository, we don't worry about it.

So simple case is to iterate through an AD group which your slurm users
belong to and add them to slurmdbd. Once they are in there, you can set
defaults with exceptions for specific users.
If you are only looking to have settings apply to all users, you don't
have to import the users. Set the QoS for the partition.

Brian Andrus

On 11/20/2023 1:45 PM, ego...@posteo.me wrote:

Hello,

I'd like to configure some sort of partition QoS so that the number of
jobs or cpus is limited for a single user.
So far my testing always depends on creating users within the
accounting database however I'd like to avoid managing each user and
having to create or sync _all_ LDAP users also within Sturm.
Or - are there solutions to sync LDAP or AzureAD users to the Slurm
accounting database?

Thanks for any input.


Best - Eg.









[slurm-users] Slurm version 23.11 is now available

2023-11-21 Thread Tim Wickberg

We are pleased to announce the availability of the Slurm 23.11 release.

To highlight some new features in 23.11:

- Substantially overhauled the SlurmDBD association management code. For 
clusters updated to 23.11, account and user additions or removals are 
significantly faster than in prior releases.


- Overhauled 'scontrol reconfigure' to prevent configuration mistakes 
from disabling slurmctld / slurmd. Instead, an error will be returned, 
and the running configuration will persist. This does require updates to 
the systemd service files to use the --systemd option to slurmctld / slurmd.


- Added a new internal auth/cred plugin - "auth/slurm". This builds off 
the prior auth/jwt model, and permits operation of the slurmdbd and 
slurmctld without access to full directory information with a suitable 
configuration.


- Added a new --external-launcher option to srun, which is automatically 
set by common MPI launcher implementations and ensures processes using 
those non-srun launchers have full access to all resources allocated on 
each node.


- Reworked the dynamic/cloud modes of operation to allow for "fanout" - 
where Slurm communication can be automatically offloaded to compute 
nodes for increased cluster scalability.


- Added initial official Debian packaging support.

- Overhauled and extended the Reservation subsystem to allow for most of 
the same resource requirements as are placed on the job. Notably, this 
permits reservations to now reserve GRES directly.


The "Slurm 23.02, 23.11, and Beyond" presentation from the Slurm 
Community BoF at SC23 (https://slurm.schedmd.com/publications.html) has 
an overview of this release.


The Slurm documentation at https://slurm.schedmd.com has also been 
updated to the 23.11 release. (Older versions can be found in the 
archive, linked from the main documentation page.)


Slurm can be downloaded from https://www.schedmd.com/downloads.php .

- Tim

--
Tim Wickberg
Chief Technology Officer, SchedMD LLC
Commercial Slurm Development and Support