Sorry, obvs wasn’t ready to send that last message yet…

Our issue is the shared storage is via NFS, and the “fast storage in limited 
supply” is only local on each node. Hence the need to copy it over from NFS 
(and then remove it when finished with it.)

I also wanted the copy & remove to be different jobs, because the main 
processing job usually requires GPU gres, which is a time-limited resource on 
the partition. I don’t want to tie up the allocation of GPUs while the data is 
staged (and removed), and if the data copy fails, don’t want to even progress 
to the job where the compute happens (so like, copy_data_locally && 
process_data)


If you've got other fast storage in limited supply that can be used for data 
that can be staged, then by all means use it, but consider whether you want 
batch cpu cores tied up with the wall time of transferring the data. This could 
easily be done on a time-shared frontend login node from which the users could 
then submit (via script) jobs after the data was staged. Most of the transfer 
wallclock is in network wait, so don't waste dedicated cores for it.

Reply via email to