On 15/9/18 2:45 am, Keith Ball wrote:
So we figured out the problem with "slurmd -C": we had run rpmbuild
on the POWER9 node, but did not have the hwloc-package installed. The
build process looks for this, and if not found, will apparently note
use hwloc/lstopo even if installed post-build.
Co
So we figured out the problem with "slurmd -C": we had run rpmbuild on the
POWER9 node, but did not have the hwloc-package installed. The build
process looks for this, and if not found, will apparently note use
hwloc/lstopo even if installed post-build.
Now Slurm reports the expected topology for
Chris,
> > 1.) Slurm seems to be incapable of recognizing sockets/cores/threads on
> > these systems.
> [...]
> > Anyone know if there is a way to get Slurm to recognize the true
topology
> > for POWER nodes?
>
> IIIRC Slurm uses hwloc for discovering topology, so "lstopo-no-graphics"
might
> give
Hi Keith,
On Tuesday, 11 September 2018 7:46:14 AM AEST Keith Ball wrote:
> 1.) Slurm seems to be incapable of recognizing sockets/cores/threads on
> these systems.
[...]
> Anyone know if there is a way to get Slurm to recognize the true topology
> for POWER nodes?
IIIRC Slurm uses hwloc for dis
Hi All,
We have installed slurm 17.11.8 on IBM AC922 nodes (POWER9) that have 4
GPUs each, and are running RHEL 7.5-ALT. Physically, these are 2-socket
nodes, with each socket having 20 cores. Depending on SMT setting (SMT1,
SMT2, SMT4) there can be 40, 80, or 160 "processors/CPUs" virtually.
Som