On Fri, 8 Jun 2012 at 5:06pm, Bill Broadley wrote > Do you think it's worth bundling up for others to use? > > This is how it works: > 1) User runs publish <directory> <name> before they start submitting > jobs. > 2) The publish command makes a torrent of that directory and starts > seeding that torrent. > 3) The user submits an arbitrary number of jobs that needs that > directory. Inside the job they "$ subscribe <name>" > 4) The subscribe command launches one torrent client per node (not per j > job) and blocks until the directory is completely downloaded > 5) /scratch/<user>/<name> has the users data > > Not nearly as convenient as having a fast parallel filesystem, but seems > potentially useful for those who have large read only datasets, GigE and > NFS. > > Thoughts?
I would definitely be interested in a tool like this. Our situation is about as you describe -- we don't have the budget or workload to justify any interconnect higher-end than GigE, but have folks who pound our central storage to get at DBs stored there. -- Joshua Baker-LePain QB3 Shared Cluster Sysadmin UCSF _______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf