Due to needing to support existing HPC workflows. I have a need to pass a
bash script within a python subprocess. It was working great with openpbs,
now I need to convert it to SLURM. I have it largely working in SLURM
hosted on Ubuntu 20.04 except that the job array is not being populated.

I've read from another user that BASH may try to evaluate variables before
they are defined by the SLURM job. I've also seen that errors in SBATCH
directives, such as a non-alphanumeric job name, can cause SLURM to not
evaluate the following directives. Can someone advise me on when SLURM
populates variables?

I have a StackOverflow post here
https://stackoverflow.com/questions/74323372/slurm-array-job-bash-scripting-within-python-subprocess

Regards,
-- 

*Chase Schuette **Pronouns: He/Him/His** | Caterpillar*

*Autonomy High Performance Computing | Iowa State University Relations*

Mobile: 507-475-1949 | Email: chase.schue...@gmail.com
<chaseschue...@gmail.com> | LinkedIn
<https://www.linkedin.com/in/chaseschuette/>
Schedule 15mins here: https://calendly.com/chaseschuette

Reply via email to