You could set /etc/security/limits.conf on every node to contain something like 
(check my syntax):

* soft core 0
* hard  core 0

And make sure that /etc/pam.d/slurm.* and /etc/pam.d/system-auth* contain:

session     required      pam_limits.so
session     required      pam_limits.so

…so that limits are enforced for each user session. We have these lines in 
several other PAM files, but those above might be the minimum set for use with 
SLURM and SSH. Both sets of files might not be necessary, but if you allow ssh 
to compute nodes after a job is started, you probably need both.

Best,
Bill.


-- 
Bill Barth, Ph.D., Director, HPC
bba...@tacc.utexas.edu        |   Phone: (512) 232-7069
Office: ROC 1.435            |   Fax:   (512) 475-9445
 
 

On 3/21/18, 6:08 AM, "slurm-users on behalf of Ole Holm Nielsen" 
<slurm-users-boun...@lists.schedmd.com on behalf of ole.h.niel...@fysik.dtu.dk> 
wrote:

    We experience problems with MPI jobs dumping lots (1 per MPI task) of 
    multi-GB core dump files, causing problems for file servers and compute 
    nodes.
    
    The user has "ulimit -c 0" in his .bashrc file, but that's ignored when 
    slurmd starts the job, and the slurmd process limits are employed in stead.
    
    I should mention that we have decided to configure slurm.conf with
       PropagateResourceLimitsExcept=ALL
    because it's desirable to have rather restrictive user limits on login 
    nodes.  Unfortunately, this means that the user's "ulimit -c 0" isn't 
    propagated to any batch job.
    
    What's the best way to suppress core dump files from jobs?  Does anyone 
    have good or bad experiences?
    
    One working solution is to modify the slurmd Systemd service file 
    /usr/lib/systemd/system/slurmd.service to add a line:
       LimitCORE=0
    I've documented further details in my Slurm Wiki page 
    
https://wiki.fysik.dtu.dk/niflheim/Slurm_configuration#slurmd-systemd-limits. 
      However, it's a bit cumbersome to modify the Systemd service file on 
    all compute nodes.
    
    Thanks for sharing any experiences.
    
    /Ole
    
    

Reply via email to