searches. Without array task scheduling this would require 500,000 individual job submissions. The fact that I never met a serious PBS shop that had not

what's wrong with 500k job submissions?  to me, the existence of "array jobs"
is an admission that the job/queueing system is inefficient. if you're saying that the issue is not per-job overhead of submission, but rather that jobs are too short, well, I think that's a user problem. I think it's entirely reasonable to require user jobs to consume some minimum cpu time
(say, few minutes).

- Policy and resource allocation features are very important to people deploying these systems

so I'm curious what that means.  things like "dept A needs to be guaranteed
N cpus, but dept B gets to use whatever is left over"? or node choice based on amount of free disk? I don't really see why these sorts of issues
would be less important to more parallel environments.

- Storage speed is often more important than network speed or latency in many cases

which makes me wonder: do bio types consider using map-reduce-like
frameworks?  that is, basically distributing the work to the data.
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to