On 10/25/19 12:49 AM, Benjamin Redling wrote:
Hello everybody,
confusing:
https://slurm.schedmd.com/archive/slurm-18.08.8/news.html
"
RELEASE NOTES FOR SLURM VERSION 19.05
28 May 2019
...
"
This is fixed now.
Bug-tracking is only via commercial support?
Anyone is welcome to file issues.
Jurgen,
Thank you for all of the information. I appreciate you taking the time to
test the configuration with 19.05.
I feel a little better about my efforts now. :-)
I will check out some of your suggestions to mitigate the issue.
Ultimately, we will probably use cgroups for containment but as I
Mark,
Thanks for responding.
Yes, it will constrain it to the amount of memory the user asked for. In
fact I have gotten that to work.
That is not the behavior that we desire (at least initially). The test
code I ran through (which just allocates chunks of RAM in a loop)
would be *constrained
Hi Mike,
IIRC, I once did some tests with the very same configuration as
your's, i.e. `JobAcctGatherType=jobacct_gather/linux´ and
`JobAcctGatherParams=OverMemoryKill´ and got this to work as expected:
Jobs were killed when they exceeded the requested amount of memory.
This was with Slurm 18.08.7.
You might want to look at these options for SchedulerParamtesr:
*preempt_strict_order*
If set, then execute extra logic in an attempt to preempt only the
lowest priority jobs. It may be desirable to set this configuration
parameter when there are multiple priorities of preemptable jobs.
need. I simply want to enforce the memory limits as specified by the user
at job submission time. This seems to have been the behavior in previous
but cgroups (with Constrain) do that all by themselves.
If someone could post just a simple slurm.conf file that forces the memory
limits to be
Hi;
The Slurm documentation at these pages:
https://slurm.schedmd.com/slurm.conf.html
https://slurm.schedmd.com/cons_res_share.html
conflict with the slurm 19.05 release notes at this page:
https://slurm.schedmd.com/news.html
Probably the documentation pages are obsolete. But, I don't know
Ahmet,
Thank you for taking the time to respond to my question.
Yes, the --mem=1GBB is a typo. It's correct in my script, I just
fat-fingered it in the email. :-)
BTW, the exact version I am using is 19.05.*2.*
Regarding your response, it seems that that might be more than what I
need. I si
Aren't we here talking about the job accounting gather plugin?
I would say, you have to use cgroups, together with
ProctrackType=proctrack/cgroup and TaskPlugin=task/cgroup if you want ot
use the jobacct_gather/cgroup plugin, because elsewise SLURMdoes not
pack the jobs into cgroups.
Best
Ma
Hi,
Let's say I have two partitions assigned to the same single load in the cluster.
LowPrio with PreemptMode=suspend Priority=1
HighPrio with PreemtMode=off Priority=5
I have 4 identical jobs requiring 1/2 of resources in the cluster. So only 2
jobs can run at the same time.
I queue up two of t
10 matches
Mail list logo