Re: [slurm-users] Gentle memory limits in Slurm using cgroup?

2019-05-02 Thread Janne Blomqvist
On 02/05/2019 17.53, Alexander Åhman wrote: > Hi, > Is it possible to configure slurm/cgroups in such way that jobs that are > using more memory than they asked for are not killed if there still are > free memory available on the compute node? When free memory gets low > these jobs can be killed as

[slurm-users] Gentle memory limits in Slurm using cgroup?

2019-05-02 Thread Alexander Åhman
Hi, Is it possible to configure slurm/cgroups in such way that jobs that are using more memory than they asked for are not killed if there still are free memory available on the compute node? When free memory gets low these jobs can be killed as usual. Today when a job has exceeded its limits

Re: [slurm-users] How should I do so that jobs are allocated to the thread and not to the core ?

2019-05-02 Thread Doug Meyer
Had same problem in slurm 15, not sure if it affects newer versions. Don’t use the expanded node definition NodeName = DEFAULT Boards = 1 SocketsPerBoard = 2 CoresPerSocket = 18 ThreadsPerCore = 2 RealMemory = 128000 Use the simpler NodeName=DEFAULT Cores=36 RealMemory = 128000 Slurm will us

Re: [slurm-users] job startup timeouts?

2019-05-02 Thread Riebs, Andy
This proved to be a scaling problem in PMIX; thanks to Artem Polyakov for tracking this down (and submitting a fix). Thanks for all the suggestions folks! Andy From: Riebs, Andy Sent: Friday, April 26, 2019 11:24 AM To: slurm-users@lists.schedmd.c

Re: [slurm-users] How should I do so that jobs are allocated to the thread and not to the core ?

2019-05-02 Thread Benjamin Redling
Have you Seen the slurm FAQ? You may want to search on that site for "Hyperthreading" (Sorry for the TOFU. vacation, mobile) Am 30. April 2019 18:07:03 MESZ schrieb Jean-mathieu CHANTREIN : >Hello, > >Most jobs of my users are single-thread. I have multithreaded >processors. The jobs seem to re