Dang it. That's it. I recently changed the default time limit on some of
my partitions, to only 48 hours. I have a reservation that starts on
Friday at 5 PM. These jobs are all assigned to partitions that still
have longer time limits. I forgot that not all partitions have the new
48-hour limit.
Still, Slurm should provide a better error message for that situation,
since I'm sure it's not that uncommon for this to happen. It would
certainly result in a lot less tickets being sent to me.
Prentice Bisbal
Lead Software Engineer
Princeton Plasma Physics Laboratory
http://www.pppl.gov
On 05/07/2018 05:11 PM, Ryan Novosielski wrote:
In my experience, it may say that even if it has nothing to do with the reason
the job isn’t running, if there are nodes on the system that aren’t available.
I assume you’ve checked for reservations?
On May 7, 2018, at 5:06 PM, Prentice Bisbal <pbis...@pppl.gov> wrote:
Dear Slurm Users,
On my cluster, I have several partitions, each with their own QOS, time limits,
etc.
Several times today, I've received complaints from users that they submitted jobs to a
partition with available nodes, but jobs are stuck in the PD state. I have spent the
majority of my day investigating this, but haven't turned up anything meaningful. Both
jobs show the "ReqNodeNotAvail" reason, but none of the nodes listed at not
available are even in the partition these jobs are submitted to. Neither job has
requested a specific node, either.
I have checked slurmctld.log on the server, and have not been able to find any
clues. Any where else I should look? Any ideas what could be causing this?
--
____
|| \\UTGERS, |---------------------------*O*---------------------------
||_// the State | Ryan Novosielski - novos...@rutgers.edu
|| \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus
|| \\ of NJ | Office of Advanced Research Computing - MSB C630, Newark
`'