Hi Matt,
Matt Hohmeister writes:
> Relatively new to Slurm here; I have someone who has asked if the
> following is possible:
>
> Allow Slurm to use as much memory on a node as exists on the node
> itself. If someone is running a process outside of Slurm, decrease
> Slurm’s memory usage to make
Relatively new to Slurm here; I have someone who has asked if the following is
possible:
Allow Slurm to use as much memory on a node as exists on the node itself. If
someone is running a process outside of Slurm, decrease Slurm's memory usage to
make way for the non-Slurm process.
Is such a th
As always, I forgot the attached file. :)
From: Vienne, Jerome
Sent: Tuesday, June 12, 2018 12:31 PM
To: 'slurm-users@lists.schedmd.com'
Subject: Proposed changes to pam_slurm_adopt
Hi,
While testing pam_slurm_adopt with multiple jobs running on the node with
Centos7, I realized that it was f
Hi,
While testing pam_slurm_adopt with multiple jobs running on the node with
Centos7, I realized that it was failing with error messages like:
Jun 12 11:57:37 server pam_slurm_adopt[86395]: From 192.168.1.48 port 36512 as
test1234: unable to determine source job
Jun 12 11:57:37 server pam_slurm
how do you plan to collect all of the performance data?
On Tue, Jun 12, 2018 at 12:06 PM, Hanby, Mike wrote:
> Howdy,
>
>
>
> Is anyone aware of any existing job completion email scripts that provide a
> summary of the jobs resource utilization? For example, something like:
>
>
>
> Job ID: 123456
Howdy,
Is anyone aware of any existing job completion email scripts that provide a
summary of the jobs resource utilization? For example, something like:
Job ID: 123456
Cluster: HPC
User/Group: jdoe/jdoe
State: COMPLETED (exit code 0)
Cores: 1
CPU Utilization: 00:18:45
CPU Efficiency: 98.60% of
Yes, I saw the same issue. Default for unset DefMemPerCPU changed from
unlimited in earlier versions to 0. I just set it to 384 in slurm.conf
so simple things run fine and make sure users always set a sane value
on submission.
On Mon, Jun 11, 2018 at 6:40 PM, Roberts, John E. wrote:
> I see this
Hi Nadav,
On Tue, Jun 12, 2018 at 8:18 AM, Nadav Toledo
wrote:
> How can one send a few jobs running in parallel with different cpus
> allocation on the same node?
According to https://slurm.schedmd.com/srun.html#OPT_cpu-bind, you may
want to use "srun --exclusive":
By default, a job step h