On 23/3/20 8:32 am, CB wrote:
I've looked at the heterogeneous job support but it creates two-separate
jobs.
Yes, but the web page does say:
# By default, the applications launched by a single execution of
# the srun command (even for different components of the
# heterogeneous job) are combi
Hi Michael,
Thanks for the comment.
I was just checking if there is any other way to do the job before
introducing another partition.
So it appears to me that creating a new partition is the way to go.
Thanks,
Chansup
On Mon, Mar 23, 2020 at 1:25 PM Renfro, Michael wrote:
> Others might have
Others might have more ideas, but anything I can think of would require a lot
of manual steps to avoid mutual interference with jobs in the other partitions
(allocating resources for a dummy job in the other partition, modifying the MPI
host list to include nodes in the other partition, etc.).
Hi Andy,
Yes, they are on teh same network fabric.
Sure, creating another partition that encompass all of the nodes of the two
or more partitions would solve the problem.
I am wondering if there are any other ways instead of creating a new
partition?
Thanks,
Chansup
On Mon, Mar 23, 2020 at 11:
When you say “distinct compute nodes,” are they at least on the same network
fabric?
If so, the first thing I’d try would be to create a new partition that
encompasses all of the nodes of the other two partitions.
Andy
From: slurm-users [mailto:slurm-users-boun...@lists.schedmd.com] On Behalf