Looks interesting, thank you. Jeff Friedman Sales Engineer o: 425.420.1291 c: 206.819.2824 www.siliconmechanics.com
On Feb 23, 2017, at 1:14 PM, Chris Dagdigian <d...@sonsorol.org> wrote: On Amazon you should be looking at CfnCluster: https://aws.amazon.com/hpc/cfncluster/ <https://aws.amazon.com/hpc/cfncluster/> The entire HPC stack is: - Directly supported by AWS - Written, used and deployed as a CloudFormation template - Very easy to extend and customize - Is under more active development than MIT StarCluster which is what we used to use for similar purposes Normally we use CfnCluster for AWS-only HPC but it should be amenable to a cloudbursting scenario especially if your VPN/VPC links are solid Chris > Jeff Friedman <mailto:jeff.fried...@siliconmechanics.com > <mailto:jeff.fried...@siliconmechanics.com>> > February 23, 2017 at 3:56 PM > Thank you all for the info, it is very useful. It seems most of the cloud > orchestrator software includes a bit more functionality than we need. We want > to use the standard HPC provisioning, scheduling, and monitoring software, > and just automate the setup and presentation of the cloud nodes. We are > looking into establishing a VPN to AWS, and then continuing to see what > software would do the best job of the automated setup/teardown of cloud > resources. We are looking at just using AWS CloudFormation as an option. > There is also Bright Computing Cluster Manager, Cycle Computing, RightScale, > and a couple others. But again, I think these are a bit to robust for what we > need. I’ll keep y’all posted if interested. > > Thanks again! > > Jeff Friedman > Sales Engineer > o: 425.420.1291 > c: 206.819.2824 > www.siliconmechanics.com > > > On Feb 23, 2017, at 10:49 AM, Lev Lafayette<lev.lafaye...@unimelb.edu.au> > wrote: > > On Wed, 2017-02-22 at 10:02 +1100, Christopher Samuel wrote: >> On 21/02/17 12:40, Lachlan Musicman wrote: >> >>> I know that it's been done successfully here by the University of >>> Melbourne's Research Platforms team - but they are bursting into the non >>> commercial Aust govt Open Stack installation Nectar. > > In context that was after (a) small test cases of cloud bursting worked > and (b) cloud bursting was used to replace our existing cloud partition. > >> So now they just provision extra VM's when they need more and add them >> to Slurm and given demand doesn't seem to go down there hasn't been a >> need to take any away yet. :-) > > Watch this space ;) > >> So this doesn't really reflect what Jeff was asking about as it's all >> the same infrastructure, it's not hitting remote clouds where you have >> to figure out how you are going to see your filesystem there, or how to >> stage data. >> > > Very much so. The ability to set up an additional partition to external > providers (e.g., amazon, azure, any openstack provider) is much less of a > problem that the interconnect issues which are quite significant. > > > All the best, > > > Lev Lafayette <mailto:lev.lafaye...@unimelb.edu.au > <mailto:lev.lafaye...@unimelb.edu.au>> > February 23, 2017 at 1:49 PM > On Wed, 2017-02-22 at 10:02 +1100, Christopher Samuel wrote: >> On 21/02/17 12:40, Lachlan Musicman wrote: >> >>> I know that it's been done successfully here by the University of >>> Melbourne's Research Platforms team - but they are bursting into the non >>> commercial Aust govt Open Stack installation Nectar. > > In context that was after (a) small test cases of cloud bursting worked > and (b) cloud bursting was used to replace our existing cloud partition. > >> So now they just provision extra VM's when they need more and add them >> to Slurm and given demand doesn't seem to go down there hasn't been a >> need to take any away yet. :-) > > Watch this space ;) > >> So this doesn't really reflect what Jeff was asking about as it's all >> the same infrastructure, it's not hitting remote clouds where you have >> to figure out how you are going to see your filesystem there, or how to >> stage data. >> > > Very much so. The ability to set up an additional partition to external > providers (e.g., amazon, azure, any openstack provider) is much less of a > problem that the interconnect issues which are quite significant. > > > All the best, > > > Christopher Samuel <mailto:sam...@unimelb.edu.au > <mailto:sam...@unimelb.edu.au>> > February 21, 2017 at 6:02 PM > > We (I help them with this) did try cloudbursting into the Melbourne Uni > Openstack instance (the same place that provided the VM's for most of > Spartan) but had to give up on it because of a bug in Slurm (since > fixed) and the unreliability of bringing up VM's - from memory we had > one particular case where it tried to boot about 50 nodes and about 20 > of them failed to start. > > So now they just provision extra VM's when they need more and add them > to Slurm and given demand doesn't seem to go down there hasn't been a > need to take any away yet. :-) > > So this doesn't really reflect what Jeff was asking about as it's all > the same infrastructure, it's not hitting remote clouds where you have > to figure out how you are going to see your filesystem there, or how to > stage data. > > cheers! > Chris > Lachlan Musicman <mailto:data...@gmail.com <mailto:data...@gmail.com>> > February 20, 2017 at 8:40 PM > I know that it's been done successfully here by the University of Melbourne's > Research Platforms team - but they are bursting into the non commercial Aust > govt Open Stack installation Nectar. http://nectar.org.au > <http://nectar.org.au/> > > cheers > L. > > > > ------ > The most dangerous phrase in the language is, "We've always done it this way." > > - Grace Hopper > > > _______________________________________________ > Beowulf mailing list, Beowulf@beowulf.org <mailto:Beowulf@beowulf.org> > sponsored by Penguin Computing > To change your subscription (digest mode or unsubscribe) visit > http://www.beowulf.org/mailman/listinfo/beowulf > <http://www.beowulf.org/mailman/listinfo/beowulf> > Jeff Friedman <mailto:jeff.fried...@siliconmechanics.com > <mailto:jeff.fried...@siliconmechanics.com>> > February 20, 2017 at 7:03 PM > Hi all, > > Has anyone dealt with bursting from an on-prem HPC cluster to the cloud > before? Are there any providers that stand out in this category? Quick > searches reveal the usual suspects (AWS, Google, etc). Just wondered what > real world experience has to say… :) > > Thanks! > > Jeff Friedman > Sales Engineer > c: 206.819.2824 > www.siliconmechanics.com <http://www.siliconmechanics.com/> > > > _______________________________________________ > Beowulf mailing list, Beowulf@beowulf.org <mailto:Beowulf@beowulf.org> > sponsored by Penguin Computing > To change your subscription (digest mode or unsubscribe) visit > http://www.beowulf.org/mailman/listinfo/beowulf > <http://www.beowulf.org/mailman/listinfo/beowulf>
_______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf