Re: [slurm-users] Using cgroups to hide GPUs on a shared controller/node

2019-05-26 Thread Steven Dick
What operating system are you running? Modern versions of systemd automatically put login sessions into their own cgroup which are themselves in a "user" group. When slurm is running parallel to this, it makes its own slurm cgroup. It should be possible to have something at boot modify the systemd

Re: [slurm-users] Using cgroups to hide GPUs on a shared controller/node

2019-05-21 Thread John Hearns
Sorry Dave, nothing handy. However look at this writeup from You Know Who: https://pbspro.atlassian.net/wiki/spaces/PD/pages/11599882/PP-325+Support+Cgroups Look at the devices: Subsystem You will need the major device number for the Nvidia devices, for example on my system: crw-rw-rw- 1 root roo

Re: [slurm-users] Using cgroups to hide GPUs on a shared controller/node

2019-05-20 Thread Dave Evans
Do you have that resource handy? I looked into the cgroups documentation but I see very little on tutorials for modifying the permissions. On Mon, May 20, 2019 at 2:45 AM John Hearns wrote: > Two replies here. > First off for normal user logins you can direct them into a cgroup - I > looked into

Re: [slurm-users] Using cgroups to hide GPUs on a shared controller/node

2019-05-20 Thread John Hearns
Two replies here. First off for normal user logins you can direct them into a cgroup - I looked into this about a year ago and it was actually quite easy. As I remember there is a service or utility available which does just that. Of course the user cgroup would not have Expanding on my theme, it

Re: [slurm-users] Using cgroups to hide GPUs on a shared controller/node

2019-05-20 Thread Nathan Harper
This doesn't directly answer your question, but in Feb last year on the ML there was a discussion about limiting user resources on login node (Stopping compute usage on login nodes).Some of the suggestions included the use of cgroups to do so, and it's possible that those methods could be exten