Thanks for the quick reply, Erick.

To follow up: 

“
Well, first you can explicitly set legacyCloud=true by using the
    Collections API CLUSTERPROP command. I don't recommend this, mind you,
    as legacyCloud will not be supported forever.
“

Yes, but like you say: we’ll have to deal with at some point, not much benefit 
in punting.

“
I'm not following something here though. When you say:
    "The desired final state of a such a deployment is a fully configured
    cluster ready to accept updates."
    are there any documents already in the index or is this really a new 
collection?
“

It’s a brand new collection with new configuration on fresh hardware which 
we’ll then fully index from a source document store (we do this when we have 
certain schema changes that require re-indexing or we want to experiment).

“
    Not sure what you mean here. Configuration of what?  Just spinning up
    a Solr node pointing to the right ZooKeeper should be sufficient, or
    I'm not understanding at all.
“

Apologies, the way I stated that was all wrong: by “requires configuration” I 
just meant to note the need to specify a shard and a node when adding a replica 
(and not even the node as you point out to me below ☺).

“
I suspect you're really talking about the "node" parameter
    to ADDREPLCIA
“

Ah, yes: that is what I meant, sorry.

It sounds like I haven’t missed too much in the documentation then, I’ll look 
more into replica placement rules.  

Thank you so much again for your time and help.

Marko


On 10/23/17, 4:33 PM, "Erick Erickson" <erickerick...@gmail.com> wrote:

    Well, first you can explicitly set legacyCloud=true by using the
    Collections API CLUSTERPROP command. I don't recommend this, mind you,
    as legacyCloud will not be supported forever.
    
    I'm not following something here though. When you say:
    "The desired final state of a such a deployment is a fully configured
    cluster ready to accept updates."
    are there any documents already in the index or is this really a new 
collection?
    
    and "adding new nodes requires explicit configuration"
    
    Not sure what you mean here. Configuration of what?  Just spinning up
    a Solr node pointing to the right ZooKeeper should be sufficient, or
    I'm not understanding at all.
    
    If not, your proposed outline seems right with one difference:
    "if a node needs to be added: provision a machine, start up Solr, use
    ADDREPLICA from Collections API passing shard number and coreNodeName"
    
    coreNodeName isn't something you ordinarily need to bother with. I'm
    being specific here where coreNodeName is usually something like
    core_node7. I suspect you're really talking about the "node" parameter
    to ADDREPLCIA, something like: 192.168.1.32:8983_solr, the entry from
    live_nodes.
    
    Now, all that said you may be better off just letting Solr add the
    replica where it wants, it'll usually put a new replica on a node
    without replicas so specifying the collection and shard should be
    sufficient. Also, note that there are replica placement rules that can
    help enforce this kind of thing.
    
    Best,
    Erick
    
    On Mon, Oct 23, 2017 at 3:12 PM, Marko Babic <babma...@abebooks.com> wrote:
    > Hi everyone,
    >
    > I'm working on upgrading a set of clusters from Solr 4.10.4 to Solr 7.1.0.
    >
    > Our deployment tooling no longer works given that legacyCloud defaults to 
false (SOLR-8256) and I'm hoping to get some advice on what to do going forward.
    >
    > Our setup is as follows:
    >   * we run in AWS with multiple independent Solr clusters, each with its 
own Zookeeper tier
    >   * each cluster hosts only a single collection
    >   * each machine/node in the cluster has a single core / is a replica for 
one shard in the collection
    >
    > We bring up new clusters as needed.  This is entirely automated and 
basically works as follows:
    >   * we first provision and set up a fresh Zookeeper tier
    >   * then, we provision a Solr bootstrapper machine that uploads 
collection config, specifies numShards and starts up
    >   * it's then easy provision the rest of the machines and have them 
automatically join a shard in the collection by hooking them to the right 
Zookeeper cluster and specifying numShards
    >   * if a node needs to be added to the cluster we just need to spin a 
machine up and start up Solr
    >
    > The desired final state of a such a deployment is a fully configured 
cluster ready to accept updates.
    >
    > Now that legacyCloud is false I'm not sure how to preserve this pretty 
nice, hands-off deployment style as the bootstrapping performed by the first 
node provisioned doesn't create a collection and adding new nodes requires 
explicit configuration.
    >
    > A new deployment procedure that I've worked out using the Collections API 
would look like:
    >   * provision Zookeeper tier
    >   * provision all the Solr nodes, wait for them all to come up
    >   * upload collection config + solr.xml to Zookeeper
    >   * create collection using Collections API
    >   * if a node needs to be added: provision a machine, start up Solr, use 
ADDREPLICA from Collections API passing shard number and coreNodeName
    >
    > This isn’t a giant deal to build but it adds complexity that I'm not 
excited about as deployment tooling needs to have some understanding of what 
the global state of the cluster is before being able to create a collection or 
when adding/replacing nodes.
    >
    > The questions I was hoping someone would have some time to help me with 
are:
    >
    > * Does the new deployment procedure I've suggested seem reasonable?  
Would we be doing anything wrong/fighting best practices?
    >   * Is there a way to keep cluster provisioning automated without having 
to build additional orchestration logic into our deployment tooling (using 
autoscaling, or triggers, or something I don’t know about)?
    >
    > Apologies for the wall of text and thanks. :)
    >
    > Marko
    >
    

Reply via email to