Re: Copying data

2020-03-17 Thread Jayadevan Maymala
On Mon, Mar 16, 2020 at 5:53 PM Erick Erickson wrote: > It’s not at all clear what the problem is. If you have a single-shard > collection, just > 1> create the stand-alone core > 2> shut down the Solr instance > 3> replace the stand-alone core's data dir with one from any of your prod > machines

Re: Copying data

2020-03-16 Thread Erick Erickson
It’s not at all clear what the problem is. If you have a single-shard collection, just 1> create the stand-alone core 2> shut down the Solr instance 3> replace the stand-alone core's data dir with one from any of your prod machines. 4> start Solr An alternative is to use the replication API t

Copying data

2020-03-15 Thread Jayadevan Maymala
Hi all, I have a 3 node Solr cluster in production (with zoo keeper). In dev, I have one node Solr instance, no zoo keeper. Which is the best way to copy over the production solr data to dev? Operating system is CentOS 7.7, Solr Version 7.3 Collection size is in the 40-50 GB range. Regards, Jayad

RE: copying data from one collection to another collection (solr cloud 521)

2015-07-15 Thread Reitzel, Charles
Alias". -Original Message- From: Raja Pothuganti [mailto:rpothuga...@competitrack.com] Sent: Wednesday, July 15, 2015 12:07 PM To: solr-user@lucene.apache.org Subject: Re: copying data from one collection to another collection (solr cloud 521) Hi Charles, Thank you for the response. We wil

Re: copying data from one collection to another collection (solr cloud 521)

2015-07-15 Thread Raja Pothuganti
;Curious if there are opportunities for optimization here. For example, >would it be faster to make a file system copy of the most recent >collection and load only changed documents (assuming the delta is >available from the source system)? > >-Original Message- >From

RE: copying data from one collection to another collection (solr cloud 521)

2015-07-15 Thread Reitzel, Charles
ld it be faster to make a file system copy of the most recent collection and load only changed documents (assuming the delta is available from the source system)? -Original Message- From: Erick Erickson [mailto:erickerick...@gmail.com] Sent: Monday, July 13, 2015 11:55 PM To: solr-user@

Re: copying data from one collection to another collection (solr cloud 521)

2015-07-13 Thread Erick Erickson
bq: does offline No. I'm talking about "collection aliasing". You can create an entirely new collection, index to it however you want then switch to using that new collection. bq: Any updates to EXISTING document in the LIVE collection should NOT be replicated to the previous week(s) snapsho

Re: copying data from one collection to another collection (solr cloud 521)

2015-07-13 Thread Raja Pothuganti
Thank you Erick >Actually, my question is why do it this way at all? Why not index >directly to your "live" nodes? This is what SolrCloud is built for. >You an use "implicit" routing to create shards say, for each week and >age out the ones that are "too old" as well. Any updates to EXISTING docu

Re: copying data from one collection to another collection (solr cloud 521)

2015-07-13 Thread Shawn Heisey
On 7/13/2015 1:49 PM, Raja Pothuganti wrote: > We are setting up a new SolrCloud environment with 5.2.1 on Ubuntu boxes. We > currently ingest data into a large collection, call it LIVE. After the full > ingest is done we then trigger a delta delta ingestion every 15 minutes to > get the documen

Re: copying data from one collection to another collection (solr cloud 521)

2015-07-13 Thread Erick Erickson
Actually, my question is why do it this way at all? Why not index directly to your "live" nodes? This is what SolrCloud is built for. There's the new backup/restore functionality that's still a work in progress, see: https://issues.apache.org/jira/browse/SOLR-5750 You an use "implicit" routing to

copying data from one collection to another collection (solr cloud 521)

2015-07-13 Thread Raja Pothuganti
Hi, We are setting up a new SolrCloud environment with 5.2.1 on Ubuntu boxes. We currently ingest data into a large collection, call it LIVE. After the full ingest is done we then trigger a delta delta ingestion every 15 minutes to get the documents & data that have changed into this LIVE insta