Hi ! Christ.
The cgroup.conf on my gpu node is as same as head node. The content are as
follow:
CgroupAutomount=yes
ConstrainCores=yes
ConstrainRAMSpace=yes
ConstrainDevices=yes
I'll try slurm of high version.
--
slurm-users mailing list -- slurm-users@lists.schedmd.com
To unsubscribe send an
Hiya,
On 4/15/25 7:03 pm, lyz--- via slurm-users wrote:
Hi, Christ. Thank you for continuing paying attention to this issue.
I followed your instuction. And This is the output:
[root@head1 ~]# systemctl cat slurmd | fgrep Delegate
Delegate=yes
That looks good to me, thanks for sharing that!
On 4/15/25 6:57 pm, lyz--- via slurm-users wrote:
Hi, Sean. It's the latest slurm version.
[root@head1 ~]# sinfo --version
slurm 22.05.3
That's quite old (and no longer supported), the oldest still supported
version is 23.11.10 and 24.11.4 came out recently.
What does the cgroup.conf file o
Hi, Christ. Thank you for continuing paying attention to this issue.
I followed your instuction. And This is the output:
[root@head1 ~]# systemctl cat slurmd | fgrep Delegate
Delegate=yes
lyz
--
slurm-users mailing list -- slurm-users@lists.schedmd.com
To unsubscribe send an email to slurm-user
Hi, Sean. It's the latest slurm version.
[root@head1 ~]# sinfo --version
slurm 22.05.3
And this is my content of the gres.conf in gpu node.
# This section of this file was automatically generated by cmd. Do not edit
manually!
# BEGIN AUTOGENERATED SECTION -- DO NOT REMOVE
Name=gpu File=/dev/nvidi
On 4/15/25 12:55 pm, Sean Crosby via slurm-users wrote:
What version of Slurm are you running and what's the contents of your
gres.conf file?
Also what does this say?
systemctl cat slurmd | fgrep Delegate
--
Chris Samuel : http://www.csamuel.org/ : Berkeley, CA, USA
--
slurm-users maili
What version of Slurm are you running and what's the contents of your gres.conf
file?
Sean
From: lyz--- via slurm-users
Sent: Tuesday, April 15, 2025 11:16:40 PM
To: slurm-users@lists.schedmd.com
Subject: [slurm-users] Re: [EXT] Re: Issue with Enforcing GPU Usa
Dear all
With the pam_slurm_adopt module as far as I understand you can ssh to a
worker node if there is at least a job running on the node by that user.
If there are multiple jobs, if I am not wrong you will be "mapped" to the
last job started on the node. And, if you are using cgroups, you will
Hi, Sean.
I followed your instructions and added ConstrainDevices=yes to the
/etc/slurm/cgroup.conf file on the server node, and then restarted the relevant
services on both the server and the client.
However, I still can't enforce the restriction in the Python program.
It seems like the restric
You need to add
ConstrainDevices=yes
To your cgroup.conf and restart slurmd on your nodes. This is the setting which
gives access to only the GRES you request in the jobs
Sean
From: lyz--- via slurm-users
Sent: Tuesday, April 15, 2025 8:29:41 PM
To: slurm-user
Hi, Christopher. Thank you for your reply.
I have already modified the cgroup.conf configuration file in Slurm as follows:
vim /etc/slurm/cgroup.conf
#
# Slurm cgroup support configuration file
#
# See man slurm.conf and man cgroup.conf for further
# information on cgroup configuration parameters
11 matches
Mail list logo