[slurm-users] sbatch problem

2024-05-28 Thread Mihai Ciubancan via slurm-users
Hello, My name is Mihai and a have an issue with a small GPU cluster manage with slurm 22.05.11. I got 2 different output when I'm trying to find out the name of the nodes(one correct and one wrong). The script is: #!/bin/bash #SBATCH --job-name=test #SBATCH --output=/data/mihai/res.txt #SBAT

[slurm-users] Re: sbatch problem

2024-05-28 Thread Mihai Ciubancan via slurm-users
ing the name of the host it is running on. If you want to avoid this, one possible solution is srun bash -c 'echo Running on host: $(hostname)' In this case the command substitution is happening after srun starts the process on a (potentially remote) node. Regards, Hermann On 5/2

[slurm-users] Re: sbatch problem

2024-05-28 Thread Mihai Ciubancan via slurm-users
ted" to srun. Which means that srun will happily run it on any (remote) node using the name of the host it is running on. If you want to avoid this, one possible solution is   srun bash -c 'echo Running on host: $(hostname)' In this case the command substitution is happening afte

[slurm-users] Re: sbatch problem

2024-05-29 Thread Mihai Ciubancan via slurm-users
on any (remote) node using the name of the host it is running on. If you want to avoid this, one possible solution is   srun bash -c 'echo Running on host: $(hostname)' In this case the command substitution is happening after srun starts the process on a (potentially remote) node. Reg

[slurm-users] jobs dropping

2024-10-25 Thread Mihai Ciubancan via slurm-users
Hello, We are trying to run some PiconGPU codes on a machine with 8x100H, susing slurm. But the jobs don't run, and are not in the queue. In slurmd logs I have: [2024-10-24T09:50:40.934] CPU_BIND: _set_batch_job_limits: Memory extracted from credential for StepId=1079.batch job_mem_limit= 64