Hi,
In our set up there are two solr clouds:
Cloud A: production cloud serves both writes and reads
Cloud B: back up cloud serves only writes
Cloud A and B have the same shard configuration.
Write requests are sent to both cloud A and B. In certain circumstances
when Cloud A's update lags behind, we want to bulk copy the binary index
from B to A.
We have tried two approaches:
Approach 1.
For cloud A:
a. delete collection to wipe out everything
b. create new collection (data is empty now)
c. shut down solr server
d. copy binary index from cloud B to corresponding shard replicas in
cloud A
e. start solr server
Approach 2.
For cloud A:
a. shut down solr server
b. remove the whole 'data' folder under index/ in each replica
c. copy binary index from cloud B to corresponding shard replicas in
cloud A
d. start solr server
Is approach 2 sufficient? I am wondering if delete/recreate collection
each time is necessary to get cloud into a "clean" state for copy binary
index between solr clouds.
Thanks for your advice!