On 17/03/2023 13:11, William Brown wrote:
We create the temporary directories using SLURM_JOB_ID, and that works
fine with Job Arrays so far as I can see. Don't you have a problem
if a user has multiple jobs on the same node?
William
Ours users just have /work/$username, anything below that
We create the temporary directories using SLURM_JOB_ID, and that works
fine with Job Arrays so far as I can see. Don't you have a problem
if a user has multiple jobs on the same node?
William
On Fri, 17 Mar 2023 at 11:17, Timo Rothenpieler
wrote:
>
> Hello!
>
> I'm currently facing a bit of an
Hello!
I'm currently facing a bit of an issue regarding cleanup after a job
completed.
I've added the following bit of Shellscript to our clusters Epilog script:
for d in "${SLURM_JOB_ID}" "${SLURM_JOB_ID}_${SLURM_ARRAY_TASK_ID}"
"${SLURM_ARRAY_JOB_ID}_${SLURM_ARRAY_TASK_ID}"; do
WO
Hi there,
we're using the jcomp_elasticsearch.c plugin (
https://github.com/SchedMD/slurm/blob/master/src/plugins/jobcomp/elasticsearch/jobcomp_elasticsearch.c)
to get some data for monitoring but we have to use the slurm-perlapi to
gather the information related to the required memory due to the
Hi Slurm Users,
first time posting. I have a new slurm setup where the users can specify
an amount of local node disk space they wish to use. This is a "gres"
resource named "local" and it measures in GB. Once the user has
scheduled a job and it gets executed, I create a folder for this job on
Dear all,
I am writing to ask you a question.
Is it possible to retrieve the status of cleared jobs (e.g. after completed
with either success or failed) from the Slurm rest interface ?
When a job (job id=131 in the example below) is cleared, the rest interface
returns this after some time after c
Message: 1
Date: Fri, 5 Mar 2021 11:56:05 +0100
From: Ole Holm Nielsen
To:
Subject: Re: [slurm-users] Get original script of a job
Message-ID: <61c47956-5fc9-5be5-9aef-08f8e27bf...@fysik.dtu.dk>
Content-Type: text/plain; charset="utf
script with each of the runs.
So I can put it in the same directory where all the other output files
go, and have a complete description of the run conditions.
----
Subject: [slurm-users] Get original script of a job
Dat
Hi,
On 5/03/2021 11:29, Alberto Morillas, Angelines wrote:
> I know that when I send a job with scontroI can get the path and the
> name of the script used to send this job, but normally the users change
> theirs scripts and sometimes all was wrong after that, so is there any
> possibility to rep
On 05-03-2021 11:29, Alberto Morillas, Angelines wrote:
I would like to know if it will be possible to get the script that was
used to send a job.
I know that when I send a job with scontroI can get the path and the
name of the script used to send this job, but normally the users change
their
Hi,
I would like to know if it will be possible to get the script that was used to
send a job.
I know that when I send a job with scontroI can get the path and the name of
the script used to send this job, but normally the users change theirs scripts
and sometimes all was wrong after that, so i
Okay ... obviously an auto-complete error that I failed to check: Please
ignore and accept my apologies.
> On Dec 16, 2019, at 7:03 AM, Wiegand, Paul wrote:
>
> unlock stokes-arcc
> get stokes-arcc
>
unlock stokes-arcc
get stokes-arcc
On 15/11/2019 17.06, Miguel Oliveira wrote:
Thank! Nice code and just what I was needing! A few wrinkles:
a) on reading the Gres from scontrol for each job on my version this is on a
TRES record not as an individual Gres. Possibly version/configuration issue.
b) converting pid2id from /proc//cg
Janne Blomqvist writes:
> On 14/11/2019 20.41, Prentice Bisbal wrote:
>> Is there any way to see how much a job used the GPU(s) on a cluster
>> using sacct or any other slurm command?
>>
>
> We have created
> https://github.com/AaltoScienceIT/ansible-role-sacct_gpu/ as a quick
> hack to put GPU uti
Thank! Nice code and just what I was needing! A few wrinkles:
a) on reading the Gres from scontrol for each job on my version this is on a
TRES record not as an individual Gres. Possibly version/configuration issue.
b) converting pid2id from /proc//cgroup is problematic on array jobs.
Again many
On 14/11/2019 20.41, Prentice Bisbal wrote:
> Is there any way to see how much a job used the GPU(s) on a cluster
> using sacct or any other slurm command?
>
We have created
https://github.com/AaltoScienceIT/ansible-role-sacct_gpu/ as a quick
hack to put GPU utilization stats into the comment fie
Do you mean akin to what some would consider "CPU efficiency" on a CPU job?
"How much... used" is a little vague.
From: slurm-users on behalf of Prentice
Bisbal
Sent: Thursday, November 14, 2019 13:41
To: Slurm User Community List
S
Is there any way to see how much a job used the GPU(s) on a cluster
using sacct or any other slurm command?
--
Prentice
Hi Jeff,
Quite close:
$ sinfo --Format=nodehost,statelong
Cheers,
--
Kilian
I use
alias sn='sinfo -Nle -o "%.20n %.15C %.8O %.7t" | uniq'
and then it's just
[root@machine]# sn
cheers
L.
--
"The antidote to apocalypticism is *apocalyptic civics*. Apocalyptic civics
is the insistence that we cannot ignore the truth, nor should we panic
about it. It is a shared consc
Subject says it all. Is there a way to get a list of nodes, their
status, and NOT have duplicate entries in the output? This is what I
have so far but it seems to duplicate nodes if they exist in more than 1
partition, which is true of all my nodes.
sinfo --Node --Format=nodelist,statelong
22 matches
Mail list logo