See: https://issues.apache.org/jira/browse/SOLR-12646
On Wed, Aug 8, 2018 at 11:24 AM, Bjarke Buur Mortensen <morten...@eluence.com> wrote: > OK, thanks. > > As long as it's my dev box, reindexing is fine. > I just hope that my assumption holds, that our prod solr is 7x segments > only. > > Thanks again, > Bjarke > > 2018-08-08 20:03 GMT+02:00 Erick Erickson <erickerick...@gmail.com>: > >> Bjarke: >> >> Using SPLITSHARD on an index with 6x segments just seems to not work, >> even outside the standalone-> cloud issue. I'll raise a JIRA. >> Meanwhile I think you'll have to re-index I'm afraid. >> >> Thanks for raising the issue. >> >> Erick >> >> On Wed, Aug 8, 2018 at 6:34 AM, Bjarke Buur Mortensen >> <morten...@eluence.com> wrote: >> > Erick, >> > >> > thanks, that is of course something I left out of the original question. >> > Our Solr is 7.1, so that should not present a problem (crossing fingers). >> > >> > However, on my dev box I'm trying out the steps, and here I have some >> > segments created with version 6 of Solr. >> > >> > After having copied data from my non-cloud solr into my >> > single-shard-single-replica collection and verified that Solr Cloud works >> > with this collection, I then submit the splitshard command >> > >> > http://172.17.0.4:8984 >> > /solr/admin/collections?action=SPLITSHARD&collection= >> procurement&shard=shard1 >> > >> > However, this gives me the error: >> > org.apache.solr.client.solrj.impl.HttpSolrClient$ >> RemoteSolrException:Error >> > from server at http://172.17.0.4:8984/solr: >> > java.lang.IllegalArgumentException: Cannot merge a segment that has been >> > created with major version 6 into this index which has been created by >> > major version 7"} >> > >> > I have tried running both optimize and IndexUpgrader on the index before >> > shard splitting, but the same error still occurs. >> > >> > Any ideas as to why this happens? >> > >> > Below is an output from running IndexUpgrader, which I cannot decipher. >> > It both states that "All segments upgraded to version 7.1.0" and ''all >> > running merges have aborted" ¯\_(ツ)_/¯ >> > >> > Thanks a lot, >> > Bjarke >> > >> > >> > ====================== >> > java -cp >> > /opt/solr/server/solr-webapp/webapp/WEB-INF/lib/lucene- >> backward-codecs-7.1.0.jar:/opt/solr/server/solr-webapp/ >> webapp/WEB-INF/lib/lucene-core-7.1.0.jar >> > org.apache.lucene.index.IndexUpgrader -delete-prior-commits -verbose >> > /var/solr/cloud/procurement_shard1_replica_n1/data/index >> > IFD 0 [2018-08-08T13:00:18.244Z; main]: init: current segments file is >> > "segments_4vs"; >> > deletionPolicy=org.apache.lucene.index.KeepOnlyLastCommitDeletionPoli >> cy@721e0f4f >> > IFD 0 [2018-08-08T13:00:18.266Z; main]: init: load commit "segments_4vs" >> > IFD 0 [2018-08-08T13:00:18.270Z; main]: now checkpoint >> > "_bhg(7.1.0):C108396" [1 segments ; isCommit = false] >> > IFD 0 [2018-08-08T13:00:18.270Z; main]: 0 msec to checkpoint >> > IW 0 [2018-08-08T13:00:18.270Z; main]: init: create=false >> > IW 0 [2018-08-08T13:00:18.273Z; main]: >> > dir=MMapDirectory@/var/solr/cloud/procurement_shard1_ >> replica_n1/data/index >> > lockFactory=org.apache.lucene.store.NativeFSLockFactory@6debcae2 >> > index=_bhg(7.1.0):C108396 >> > version=7.1.0 >> > analyzer=null >> > ramBufferSizeMB=16.0 >> > maxBufferedDocs=-1 >> > mergedSegmentWarmer=null >> > delPolicy=org.apache.lucene.index.KeepOnlyLastCommitDeletionPolicy >> > commit=null >> > openMode=CREATE_OR_APPEND >> > similarity=org.apache.lucene.search.similarities.BM25Similarity >> > mergeScheduler=ConcurrentMergeScheduler: maxThreadCount=-1, >> > maxMergeCount=-1, ioThrottle=true >> > codec=Lucene70 >> > infoStream=org.apache.lucene.util.PrintStreamInfoStream >> > mergePolicy=UpgradeIndexMergePolicy([TieredMergePolicy: >> maxMergeAtOnce=10, >> > maxMergeAtOnceExplicit=30, maxMergedSegmentMB=5120.0, floorSegmentMB=2.0, >> > forceMergeDeletesPctAllowed=10.0, segmentsPerTier=10.0, >> > maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.1) >> > indexerThreadPool=org.apache.lucene.index.DocumentsWriterPerThreadPool@ >> 5ba23b66 >> > readerPooling=true >> > perThreadHardLimitMB=1945 >> > useCompoundFile=true >> > commitOnClose=true >> > indexSort=null >> > writer=org.apache.lucene.index.IndexWriter@2ff4f00f >> > >> > IW 0 [2018-08-08T13:00:18.273Z; main]: MMapDirectory.UNMAP_SUPPORTED= >> true >> > IndexUpgrader 0 [2018-08-08T13:00:18.274Z; main]: Upgrading all pre-7.1.0 >> > segments of index directory >> > 'MMapDirectory@/var/solr/cloud/procurement_shard1_replica_n1/data/index >> > lockFactory=org.apache.lucene.store.NativeFSLockFactory@6debcae2' to >> > version 7.1.0... >> > IW 0 [2018-08-08T13:00:18.274Z; main]: forceMerge: index now >> > _bhg(7.1.0):C108396 >> > IW 0 [2018-08-08T13:00:18.274Z; main]: now flush at forceMerge >> > IW 0 [2018-08-08T13:00:18.274Z; main]: start flush: >> applyAllDeletes=true >> > IW 0 [2018-08-08T13:00:18.274Z; main]: index before flush >> > _bhg(7.1.0):C108396 >> > DW 0 [2018-08-08T13:00:18.274Z; main]: startFullFlush >> > DW 0 [2018-08-08T13:00:18.275Z; main]: main finishFullFlush success=true >> > IW 0 [2018-08-08T13:00:18.275Z; main]: now apply all deletes for all >> > segments buffered updates bytesUsed=0 reader pool bytesUsed=0 >> > BD 0 [2018-08-08T13:00:18.275Z; main]: waitApply: no deletes to apply >> > UPGMP 0 [2018-08-08T13:00:18.276Z; main]: findForcedMerges: >> > segmentsToUpgrade={} >> > MS 0 [2018-08-08T13:00:18.282Z; main]: initDynamicDefaults spins=true >> > maxThreadCount=1 maxMergeCount=6 >> > MS 0 [2018-08-08T13:00:18.282Z; main]: now merge >> > MS 0 [2018-08-08T13:00:18.282Z; main]: index: _bhg(7.1.0):C108396 >> > MS 0 [2018-08-08T13:00:18.282Z; main]: no more merges pending; now >> return >> > IndexUpgrader 0 [2018-08-08T13:00:18.282Z; main]: All segments upgraded >> to >> > version 7.1.0 >> > IndexUpgrader 0 [2018-08-08T13:00:18.283Z; main]: Enforcing commit to >> > rewrite all index metadata... >> > IW 0 [2018-08-08T13:00:18.283Z; main]: commit: start >> > IW 0 [2018-08-08T13:00:18.283Z; main]: commit: enter lock >> > IW 0 [2018-08-08T13:00:18.283Z; main]: commit: now prepare >> > IW 0 [2018-08-08T13:00:18.283Z; main]: prepareCommit: flush >> > IW 0 [2018-08-08T13:00:18.283Z; main]: index before flush >> > _bhg(7.1.0):C108396 >> > DW 0 [2018-08-08T13:00:18.283Z; main]: startFullFlush >> > IW 0 [2018-08-08T13:00:18.283Z; main]: now apply all deletes for all >> > segments buffered updates bytesUsed=0 reader pool bytesUsed=0 >> > BD 0 [2018-08-08T13:00:18.283Z; main]: waitApply: no deletes to apply >> > DW 0 [2018-08-08T13:00:18.284Z; main]: main finishFullFlush success=true >> > IW 0 [2018-08-08T13:00:18.284Z; main]: startCommit(): start >> > IW 0 [2018-08-08T13:00:18.284Z; main]: startCommit >> > index=_bhg(7.1.0):C108396 changeCount=2 >> > IW 0 [2018-08-08T13:00:18.293Z; main]: startCommit: wrote pending >> segments >> > file "pending_segments_4vt" >> > IW 0 [2018-08-08T13:00:18.295Z; main]: done all syncs: >> > [_bhg_Lucene50_0.tip, _bhg.fdx, _bhg.fnm, _bhg.nvm, _bhg.fdt, _bhg.si, >> > _bhg_Lucene50_0.pos, _bhg.nvd, _bhg_Lucene50_0.doc, _bhg_Lucene50_0.tim] >> > IW 0 [2018-08-08T13:00:18.295Z; main]: commit: pendingCommit != null >> > IW 0 [2018-08-08T13:00:18.298Z; main]: commit: done writing segments file >> > "segments_4vt" >> > IFD 0 [2018-08-08T13:00:18.298Z; main]: now checkpoint >> > "_bhg(7.1.0):C108396" [1 segments ; isCommit = true] >> > IFD 0 [2018-08-08T13:00:18.298Z; main]: deleteCommits: now decRef commit >> > "segments_4vs" >> > IFD 0 [2018-08-08T13:00:18.298Z; main]: delete [segments_4vs] >> > IFD 0 [2018-08-08T13:00:18.299Z; main]: 0 msec to checkpoint >> > IW 0 [2018-08-08T13:00:18.319Z; main]: commit: took 16.0 msec >> > IW 0 [2018-08-08T13:00:18.319Z; main]: commit: done >> > IndexUpgrader 0 [2018-08-08T13:00:18.319Z; main]: Committed upgraded >> > metadata to index. >> > IW 0 [2018-08-08T13:00:18.319Z; main]: now flush at close >> > IW 0 [2018-08-08T13:00:18.319Z; main]: start flush: >> applyAllDeletes=true >> > IW 0 [2018-08-08T13:00:18.319Z; main]: index before flush >> > _bhg(7.1.0):C108396 >> > DW 0 [2018-08-08T13:00:18.319Z; main]: startFullFlush >> > DW 0 [2018-08-08T13:00:18.320Z; main]: main finishFullFlush success=true >> > IW 0 [2018-08-08T13:00:18.320Z; main]: now apply all deletes for all >> > segments buffered updates bytesUsed=0 reader pool bytesUsed=0 >> > BD 0 [2018-08-08T13:00:18.320Z; main]: waitApply: no deletes to apply >> > MS 0 [2018-08-08T13:00:18.320Z; main]: updateMergeThreads ioThrottle=true >> > targetMBPerSec=10240.0 MB/sec >> > MS 0 [2018-08-08T13:00:18.320Z; main]: now merge >> > MS 0 [2018-08-08T13:00:18.321Z; main]: index: _bhg(7.1.0):C108396 >> > MS 0 [2018-08-08T13:00:18.321Z; main]: no more merges pending; now >> return >> > IW 0 [2018-08-08T13:00:18.321Z; main]: waitForMerges >> > IW 0 [2018-08-08T13:00:18.321Z; main]: waitForMerges done >> > IW 0 [2018-08-08T13:00:18.321Z; main]: commit: start >> > IW 0 [2018-08-08T13:00:18.321Z; main]: commit: enter lock >> > IW 0 [2018-08-08T13:00:18.321Z; main]: commit: now prepare >> > IW 0 [2018-08-08T13:00:18.321Z; main]: prepareCommit: flush >> > IW 0 [2018-08-08T13:00:18.321Z; main]: index before flush >> > _bhg(7.1.0):C108396 >> > DW 0 [2018-08-08T13:00:18.321Z; main]: startFullFlush >> > IW 0 [2018-08-08T13:00:18.321Z; main]: now apply all deletes for all >> > segments buffered updates bytesUsed=0 reader pool bytesUsed=0 >> > BD 0 [2018-08-08T13:00:18.322Z; main]: waitApply: no deletes to apply >> > DW 0 [2018-08-08T13:00:18.322Z; main]: main finishFullFlush success=true >> > IW 0 [2018-08-08T13:00:18.322Z; main]: startCommit(): start >> > IW 0 [2018-08-08T13:00:18.322Z; main]: skip startCommit(): no changes >> > pending >> > IW 0 [2018-08-08T13:00:18.322Z; main]: commit: pendingCommit == null; >> skip >> > IW 0 [2018-08-08T13:00:18.322Z; main]: commit: took 0.9 msec >> > IW 0 [2018-08-08T13:00:18.322Z; main]: commit: done >> > IW 0 [2018-08-08T13:00:18.322Z; main]: rollback >> > IW 0 [2018-08-08T13:00:18.322Z; main]: all running merges have aborted >> > IW 0 [2018-08-08T13:00:18.323Z; main]: rollback: done finish merges >> > DW 0 [2018-08-08T13:00:18.323Z; main]: abort >> > DW 0 [2018-08-08T13:00:18.323Z; main]: done abort success=true >> > IW 0 [2018-08-08T13:00:18.323Z; main]: rollback: >> infos=_bhg(7.1.0):C108396 >> > IFD 0 [2018-08-08T13:00:18.323Z; main]: now checkpoint >> > "_bhg(7.1.0):C108396" [1 segments ; isCommit = false] >> > IFD 0 [2018-08-08T13:00:18.323Z; main]: 0 msec to checkpoint >> > >> > 2018-08-07 16:38 GMT+02:00 Erick Erickson <erickerick...@gmail.com>: >> > >> >> Bjarke: >> >> >> >> One thing, what version of Solr are you moving _from_ and _to_? >> >> Solr/Lucene only guarantee one major backward revision so you can copy >> >> an index created with Solr 6 to another Solr 6 or Solr 7, but you >> >> couldn't copy an index created with Solr 5 to Solr 7... >> >> >> >> Also note that shard splitting is a very expensive operation, so be >> >> patient.... >> >> >> >> Best, >> >> Erick >> >> >> >> On Tue, Aug 7, 2018 at 6:17 AM, Rahul Singh >> >> <rahul.xavier.si...@gmail.com> wrote: >> >> > Bjarke, >> >> > >> >> > I am imagining that at some point you may need to shard that data if >> it >> >> grows. Or do you imagine this data to remain stagnant? >> >> > >> >> > Generally you want to add solrcloud to do two things : 1. Increase >> >> availability with replicas 2. Increase available data via shards 3. >> >> Increase fault tolerance with leader and replicas being spread around >> the >> >> cluster. >> >> > >> >> > You would be bypassing general High availability / distributed >> computing >> >> processes by trying to not reindex. >> >> > >> >> > Rahul >> >> > On Aug 7, 2018, 7:06 AM -0400, Bjarke Buur Mortensen < >> >> morten...@eluence.com>, wrote: >> >> >> Hi List, >> >> >> >> >> >> is there a cookbook recipe for moving an existing solr core to a solr >> >> cloud >> >> >> collection. >> >> >> >> >> >> We currently have a single machine with a large core (~150gb), and we >> >> would >> >> >> like to move to solr cloud. >> >> >> >> >> >> I haven't been able to find anything that reuses an existing index, >> so >> >> any >> >> >> pointers much appreciated. >> >> >> >> >> >> Thanks, >> >> >> Bjarke >> >> >>