[slurm-users] Discrepancy in TmpDisk values between scontrol and sinfo in Slurm24.05.3

2025-01-11 Thread Gizo Nanava via slurm-users
There seems to be an issue with the TmpDisk value reporting in Slurm 24.05.3. While the correct value is displayed using the scontrol show nodes command, sinfo appears to report an incorrect value under certain conditions. For example, the TmpDisk parameter for my compute nodes is configured as

[slurm-users] Flag OVERLAP in advanced reservation

2023-07-27 Thread Gizo Nanava
Hello, I observe strange behavior of advanced reservations having OVERLAP in their flag's list. If I create two advanced reservations on different set of nodes and a particular username is configured to only have an access to one with the flag OVERLAP, then the username can also run jobs on n

[slurm-users] Flag OVERLAP in advanced reservation

2023-07-23 Thread Gizo Nanava
Hello, I observe strange behavior of advanced reservations having OVERLAP in their flag's list. If I create two advanced reservations on different set of nodes and a particular username is configured to only have an access to one with the flag OVERLAP, then the username can also run jobs on n

Re: [slurm-users] Power saving and node weight

2023-03-01 Thread Gizo Nanava
Hello Brian, thanks a lot for the info. > > You may be able to use the alternate approach that I was able to do as well. > I would be insterested in any alternatives. Could you point me to some doc? Best wishes Gizo > Brian Andrus > > > On 2/28/2023 7:44 AM, Gizo Na

[slurm-users] Power saving and node weight

2023-02-28 Thread Gizo Nanava
Hello, it seems that if a slurm power saving is enabled then the parameter "Weight" seem to be ignored for nodes that are in a power down state. Is there any way to make the option working for a cluster running slurm in a powe saveing mode?. I am aware of the note to the weight option in the

Re: [slurm-users] job_container/tmpfs and autofs

2023-01-12 Thread Gizo Nanava
enburger Platz 1 > 13353 Berlin >   > magnus.hagd...@charite.de > https://www.charite.de > HPC Helpdesk: sc-hpc-helpd...@charite.de -- ___ Dr. Gizo Nanava Group Leader, Scientific Computing Leibniz Universität IT Services Leibniz Universität Hannover Schlosswender Str. 5 D-30159 Hannover Tel +49 511 762 7919085 http://www.luis.uni-hannover.de

Re: [slurm-users] salloc problem

2022-11-30 Thread Gizo Nanava
, presumable caused by some race conditions - in very rare cases, salloc works without this issue. I see that doc on the Slurm power saving mentions about salloc, but not for the case of interactive use of it. Thank you & best regards Gizo > On 27/10/22 4:18 am, Gizo Nanava wrote: > >

[slurm-users] salloc problem

2022-10-27 Thread Gizo Nanava
Hello, we run into another issue when using salloc interactively on a cluster where Slurm power saving is enabled. The problem seems to be caused by the job_container plugin and occurs when the job starts on a node which boots from a power down state. If I resubmit a job immediately after the

Re: [slurm-users] Slurm Power Saving & salloc

2022-10-25 Thread Gizo Nanava
Please ignore the question - the option SchedulerParameters=salloc_wait_nodes solves the issue. kind regards Gizo > Hello, > > it seems that in a cluster configured for power saving, salloc does not wait > until the nodes > assigned to the job recover from the power down state and go back

[slurm-users] Slurm Power Saving & salloc

2022-10-24 Thread Gizo Nanava
Hello, it seems that in a cluster configured for power saving, salloc does not wait until the nodes assigned to the job recover from the power down state and go back to normal operation Although the job is in the state CONFIGURING and the node are still in IDLE+NOT_RESPONDING+POWERING_UP, th

[slurm-users] salloc not working in configless setup on login machine

2020-03-03 Thread nanava
lurmctl is identified through dns srv record. If I run sinfo without prior execution of salloc, then it works. Do cluster login nodes still require slurm.conf file? Thank you. Best regards Gizo -- ___ Dr. Gizo Nanava Leibniz Universitaet IT Services Leibniz Universitaet Ha

[slurm-users] Node allocation policies in cons_res and cons_tres

2020-02-23 Thread nanava
Hi, it seems cons_res and cons_tres allocate cpus across nodes differently. The doc here https://slurm.schedmd.com/cpu_management.html#Overview "When using SelectType=select/cons_res, the default allocation method across nodes is block allocation (allocate all available CPUs in a node before u