Hi Steve,
The requirement for a client node as I tested is
* munge daemon for auth
* mechanism for client to obtain configuration
So yes I believe you would need munge working on the submitting machine.
For the configuration, I used to keep a copy of the slurm config in
/etc/slurm in the client node via NFS mount, and now I am using DNS SRV
record with "configless slurm". That simplified my setup by a lot as I
only need to install slurm client and munge, then copy the munge key to
the client machine and the client machine can just find the slurmctld,
get the configuration from there and submit job.
Hope that helps
S. Zhang
On 2023/08/28 7:10, Steven Swanson wrote:
This makes sense and I have a unified authorization scheme, but how do
the slurm commands (e.g., salloc and squeue) know which slurm head
node to talk to? Do I have to run munge on the submitting machines?
-steve
On Sun, Aug 27, 2023 at 5:00 AM
<slurm-users-requ...@lists.schedmd.com> wrote:
Message: 3
Date: Sun, 27 Aug 2023 11:10:18 +0100
From: William Brown <will...@signalbox.org.uk>
To: Slurm User Community List <slurm-users@lists.schedmd.com>
Subject: Re: [slurm-users] Submitting jobs from machines outside the
cluster
Message-ID:
<CAGdVoZirihTKz7V=yeovenkm7lcjoe-jrzdaga+h2wgbndk...@mail.gmail.com
<mailto:yeovenkm7lcjoe-jrzdaga%2bh2wgbndk...@mail.gmail.com>>
Content-Type: text/plain; charset="utf-8"
The machine that runs the CLI isn't usually a cluster node or run
slurmd.
The control node has to accept the user is who they claim and
AFAIK that is
the job of munge. And your onboard and external firewalls must
allow the
requisite ports.
We used Galaxy (see galaxy.eu <http://galaxy.eu> if unfamiliar)
and could submit jobs to
various job runners including Slurm. The galaxy node definitely
didn't run
any slurm daemons.
I think you do need a common authentication system between the
submitting
node and the cluster, but that may just be what I'm used to.
William Brown
On Sun, 27 Aug 2023, 07:20 Steven Swanson, <sjswan...@ucsd.edu> wrote:
> Can I submit jobs with a computer/docker container that is not
part of the
> slurm cluster?
>
> I'm trying to set up slurm as the backend for a system with Jupyter
> Notebook-based front end.
>
> The jupyter notebooks are running in containers managed by
Jupyter Hub,
> which is a mostly turnkey system for providing docker containers
that users
> can access via jupyter.
>
> I would like the jupyter containers to be able to submit jobs to
slurm,
> but making them part of the cluster doesn't seem to make sense
because:
>
> 1. They are dynamically created and don't have known hostnames.
> 2. They aren't supposed to run jobs.
>
> Is there a way to do this? I tried just running slurmd in the
jupyter
> containers, but it complained about not being able to figure out
its name
> (I think because the container's hostname is not listed in
slurm.conf).
>
> My fall back solution is to use ssh to connect to the slurm head
node and
> run jobs there, but that seems kludgy.
>
> -steve
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<https://urldefense.com/v3/__http://lists.schedmd.com/pipermail/slurm-users/attachments/20230827/fa37f4ae/attachment-0001.htm__;!!Mih3wA!H6jtRH1Co7St0JJ5HiauL4ay4-Eh6LH0U82KsCeTJHkkjp-ljokZ6CgLmcjQ0N_QpvgxDYbrcQv4bQYRUvStqM0vuZcLCGlWVg$
>
End of slurm-users Digest, Vol 70, Issue 36
*******************************************