Hello,
My name is Mihai and a have an issue with a small GPU cluster manage
with slurm 22.05.11. I got 2 different output when I'm trying to find
out the name of the nodes(one correct and one wrong). The script is:
#!/bin/bash
#SBATCH --job-name=test
#SBATCH --output=/data/mihai/res.txt
#SBAT
ing the name of the host it is running
on.
If you want to avoid this, one possible solution is
srun bash -c 'echo Running on host: $(hostname)'
In this case the command substitution is happening after srun starts
the process on a (potentially remote) node.
Regards,
Hermann
On 5/2
ted" to srun. Which means that srun will happily
run it on any (remote) node using the name of the host it is running
on.
If you want to avoid this, one possible solution is
srun bash -c 'echo Running on host: $(hostname)'
In this case the command substitution is happening afte
on any (remote) node using the name of the host it is running
on.
If you want to avoid this, one possible solution is
srun bash -c 'echo Running on host: $(hostname)'
In this case the command substitution is happening after srun starts
the process on a (potentially remote) node.
Reg
Hello,
We are trying to run some PiconGPU codes on a machine with 8x100H,
susing slurm. But the jobs don't run, and are not in the queue. In
slurmd logs I have:
[2024-10-24T09:50:40.934] CPU_BIND: _set_batch_job_limits: Memory
extracted from credential for StepId=1079.batch job_mem_limit= 64