OK, thanks.

As long as it's my dev box, reindexing is fine.
I just hope that my assumption holds, that our prod solr is 7x segments
only.

Thanks again,
Bjarke

2018-08-08 20:03 GMT+02:00 Erick Erickson <erickerick...@gmail.com>:

> Bjarke:
>
> Using SPLITSHARD on an index with 6x segments just seems to not work,
> even outside the standalone-> cloud issue. I'll raise a JIRA.
> Meanwhile I think you'll have to re-index I'm afraid.
>
> Thanks for raising the issue.
>
> Erick
>
> On Wed, Aug 8, 2018 at 6:34 AM, Bjarke Buur Mortensen
> <morten...@eluence.com> wrote:
> > Erick,
> >
> > thanks, that is of course something I left out of the original question.
> > Our Solr is 7.1, so that should not present a problem (crossing fingers).
> >
> > However, on my dev box I'm trying out the steps, and here I have some
> > segments created with version 6 of Solr.
> >
> > After having copied data from my non-cloud solr into my
> > single-shard-single-replica collection and verified that Solr Cloud works
> > with this collection, I then submit the splitshard command
> >
> > http://172.17.0.4:8984
> > /solr/admin/collections?action=SPLITSHARD&collection=
> procurement&shard=shard1
> >
> > However, this gives me the error:
> > org.apache.solr.client.solrj.impl.HttpSolrClient$
> RemoteSolrException:Error
> > from server at http://172.17.0.4:8984/solr:
> > java.lang.IllegalArgumentException: Cannot merge a segment that has been
> > created with major version 6 into this index which has been created by
> > major version 7"}
> >
> > I have tried running both optimize and IndexUpgrader on the index before
> > shard splitting, but the same error still occurs.
> >
> > Any ideas as to why this happens?
> >
> > Below is an output from running IndexUpgrader, which I cannot decipher.
> > It both states that "All segments upgraded to version 7.1.0" and ''all
> > running merges have aborted" ¯\_(ツ)_/¯
> >
> > Thanks a lot,
> > Bjarke
> >
> >
> > ======================
> > java -cp
> > /opt/solr/server/solr-webapp/webapp/WEB-INF/lib/lucene-
> backward-codecs-7.1.0.jar:/opt/solr/server/solr-webapp/
> webapp/WEB-INF/lib/lucene-core-7.1.0.jar
> > org.apache.lucene.index.IndexUpgrader -delete-prior-commits -verbose
> > /var/solr/cloud/procurement_shard1_replica_n1/data/index
> > IFD 0 [2018-08-08T13:00:18.244Z; main]: init: current segments file is
> > "segments_4vs";
> > deletionPolicy=org.apache.lucene.index.KeepOnlyLastCommitDeletionPoli
> cy@721e0f4f
> > IFD 0 [2018-08-08T13:00:18.266Z; main]: init: load commit "segments_4vs"
> > IFD 0 [2018-08-08T13:00:18.270Z; main]: now checkpoint
> > "_bhg(7.1.0):C108396" [1 segments ; isCommit = false]
> > IFD 0 [2018-08-08T13:00:18.270Z; main]: 0 msec to checkpoint
> > IW 0 [2018-08-08T13:00:18.270Z; main]: init: create=false
> > IW 0 [2018-08-08T13:00:18.273Z; main]:
> > dir=MMapDirectory@/var/solr/cloud/procurement_shard1_
> replica_n1/data/index
> > lockFactory=org.apache.lucene.store.NativeFSLockFactory@6debcae2
> > index=_bhg(7.1.0):C108396
> > version=7.1.0
> > analyzer=null
> > ramBufferSizeMB=16.0
> > maxBufferedDocs=-1
> > mergedSegmentWarmer=null
> > delPolicy=org.apache.lucene.index.KeepOnlyLastCommitDeletionPolicy
> > commit=null
> > openMode=CREATE_OR_APPEND
> > similarity=org.apache.lucene.search.similarities.BM25Similarity
> > mergeScheduler=ConcurrentMergeScheduler: maxThreadCount=-1,
> > maxMergeCount=-1, ioThrottle=true
> > codec=Lucene70
> > infoStream=org.apache.lucene.util.PrintStreamInfoStream
> > mergePolicy=UpgradeIndexMergePolicy([TieredMergePolicy:
> maxMergeAtOnce=10,
> > maxMergeAtOnceExplicit=30, maxMergedSegmentMB=5120.0, floorSegmentMB=2.0,
> > forceMergeDeletesPctAllowed=10.0, segmentsPerTier=10.0,
> > maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.1)
> > indexerThreadPool=org.apache.lucene.index.DocumentsWriterPerThreadPool@
> 5ba23b66
> > readerPooling=true
> > perThreadHardLimitMB=1945
> > useCompoundFile=true
> > commitOnClose=true
> > indexSort=null
> > writer=org.apache.lucene.index.IndexWriter@2ff4f00f
> >
> > IW 0 [2018-08-08T13:00:18.273Z; main]: MMapDirectory.UNMAP_SUPPORTED=
> true
> > IndexUpgrader 0 [2018-08-08T13:00:18.274Z; main]: Upgrading all pre-7.1.0
> > segments of index directory
> > 'MMapDirectory@/var/solr/cloud/procurement_shard1_replica_n1/data/index
> > lockFactory=org.apache.lucene.store.NativeFSLockFactory@6debcae2' to
> > version 7.1.0...
> > IW 0 [2018-08-08T13:00:18.274Z; main]: forceMerge: index now
> > _bhg(7.1.0):C108396
> > IW 0 [2018-08-08T13:00:18.274Z; main]: now flush at forceMerge
> > IW 0 [2018-08-08T13:00:18.274Z; main]:   start flush:
> applyAllDeletes=true
> > IW 0 [2018-08-08T13:00:18.274Z; main]:   index before flush
> > _bhg(7.1.0):C108396
> > DW 0 [2018-08-08T13:00:18.274Z; main]: startFullFlush
> > DW 0 [2018-08-08T13:00:18.275Z; main]: main finishFullFlush success=true
> > IW 0 [2018-08-08T13:00:18.275Z; main]: now apply all deletes for all
> > segments buffered updates bytesUsed=0 reader pool bytesUsed=0
> > BD 0 [2018-08-08T13:00:18.275Z; main]: waitApply: no deletes to apply
> > UPGMP 0 [2018-08-08T13:00:18.276Z; main]: findForcedMerges:
> > segmentsToUpgrade={}
> > MS 0 [2018-08-08T13:00:18.282Z; main]: initDynamicDefaults spins=true
> > maxThreadCount=1 maxMergeCount=6
> > MS 0 [2018-08-08T13:00:18.282Z; main]: now merge
> > MS 0 [2018-08-08T13:00:18.282Z; main]:   index: _bhg(7.1.0):C108396
> > MS 0 [2018-08-08T13:00:18.282Z; main]:   no more merges pending; now
> return
> > IndexUpgrader 0 [2018-08-08T13:00:18.282Z; main]: All segments upgraded
> to
> > version 7.1.0
> > IndexUpgrader 0 [2018-08-08T13:00:18.283Z; main]: Enforcing commit to
> > rewrite all index metadata...
> > IW 0 [2018-08-08T13:00:18.283Z; main]: commit: start
> > IW 0 [2018-08-08T13:00:18.283Z; main]: commit: enter lock
> > IW 0 [2018-08-08T13:00:18.283Z; main]: commit: now prepare
> > IW 0 [2018-08-08T13:00:18.283Z; main]: prepareCommit: flush
> > IW 0 [2018-08-08T13:00:18.283Z; main]:   index before flush
> > _bhg(7.1.0):C108396
> > DW 0 [2018-08-08T13:00:18.283Z; main]: startFullFlush
> > IW 0 [2018-08-08T13:00:18.283Z; main]: now apply all deletes for all
> > segments buffered updates bytesUsed=0 reader pool bytesUsed=0
> > BD 0 [2018-08-08T13:00:18.283Z; main]: waitApply: no deletes to apply
> > DW 0 [2018-08-08T13:00:18.284Z; main]: main finishFullFlush success=true
> > IW 0 [2018-08-08T13:00:18.284Z; main]: startCommit(): start
> > IW 0 [2018-08-08T13:00:18.284Z; main]: startCommit
> > index=_bhg(7.1.0):C108396 changeCount=2
> > IW 0 [2018-08-08T13:00:18.293Z; main]: startCommit: wrote pending
> segments
> > file "pending_segments_4vt"
> > IW 0 [2018-08-08T13:00:18.295Z; main]: done all syncs:
> > [_bhg_Lucene50_0.tip, _bhg.fdx, _bhg.fnm, _bhg.nvm, _bhg.fdt, _bhg.si,
> > _bhg_Lucene50_0.pos, _bhg.nvd, _bhg_Lucene50_0.doc, _bhg_Lucene50_0.tim]
> > IW 0 [2018-08-08T13:00:18.295Z; main]: commit: pendingCommit != null
> > IW 0 [2018-08-08T13:00:18.298Z; main]: commit: done writing segments file
> > "segments_4vt"
> > IFD 0 [2018-08-08T13:00:18.298Z; main]: now checkpoint
> > "_bhg(7.1.0):C108396" [1 segments ; isCommit = true]
> > IFD 0 [2018-08-08T13:00:18.298Z; main]: deleteCommits: now decRef commit
> > "segments_4vs"
> > IFD 0 [2018-08-08T13:00:18.298Z; main]: delete [segments_4vs]
> > IFD 0 [2018-08-08T13:00:18.299Z; main]: 0 msec to checkpoint
> > IW 0 [2018-08-08T13:00:18.319Z; main]: commit: took 16.0 msec
> > IW 0 [2018-08-08T13:00:18.319Z; main]: commit: done
> > IndexUpgrader 0 [2018-08-08T13:00:18.319Z; main]: Committed upgraded
> > metadata to index.
> > IW 0 [2018-08-08T13:00:18.319Z; main]: now flush at close
> > IW 0 [2018-08-08T13:00:18.319Z; main]:   start flush:
> applyAllDeletes=true
> > IW 0 [2018-08-08T13:00:18.319Z; main]:   index before flush
> > _bhg(7.1.0):C108396
> > DW 0 [2018-08-08T13:00:18.319Z; main]: startFullFlush
> > DW 0 [2018-08-08T13:00:18.320Z; main]: main finishFullFlush success=true
> > IW 0 [2018-08-08T13:00:18.320Z; main]: now apply all deletes for all
> > segments buffered updates bytesUsed=0 reader pool bytesUsed=0
> > BD 0 [2018-08-08T13:00:18.320Z; main]: waitApply: no deletes to apply
> > MS 0 [2018-08-08T13:00:18.320Z; main]: updateMergeThreads ioThrottle=true
> > targetMBPerSec=10240.0 MB/sec
> > MS 0 [2018-08-08T13:00:18.320Z; main]: now merge
> > MS 0 [2018-08-08T13:00:18.321Z; main]:   index: _bhg(7.1.0):C108396
> > MS 0 [2018-08-08T13:00:18.321Z; main]:   no more merges pending; now
> return
> > IW 0 [2018-08-08T13:00:18.321Z; main]: waitForMerges
> > IW 0 [2018-08-08T13:00:18.321Z; main]: waitForMerges done
> > IW 0 [2018-08-08T13:00:18.321Z; main]: commit: start
> > IW 0 [2018-08-08T13:00:18.321Z; main]: commit: enter lock
> > IW 0 [2018-08-08T13:00:18.321Z; main]: commit: now prepare
> > IW 0 [2018-08-08T13:00:18.321Z; main]: prepareCommit: flush
> > IW 0 [2018-08-08T13:00:18.321Z; main]:   index before flush
> > _bhg(7.1.0):C108396
> > DW 0 [2018-08-08T13:00:18.321Z; main]: startFullFlush
> > IW 0 [2018-08-08T13:00:18.321Z; main]: now apply all deletes for all
> > segments buffered updates bytesUsed=0 reader pool bytesUsed=0
> > BD 0 [2018-08-08T13:00:18.322Z; main]: waitApply: no deletes to apply
> > DW 0 [2018-08-08T13:00:18.322Z; main]: main finishFullFlush success=true
> > IW 0 [2018-08-08T13:00:18.322Z; main]: startCommit(): start
> > IW 0 [2018-08-08T13:00:18.322Z; main]:   skip startCommit(): no changes
> > pending
> > IW 0 [2018-08-08T13:00:18.322Z; main]: commit: pendingCommit == null;
> skip
> > IW 0 [2018-08-08T13:00:18.322Z; main]: commit: took 0.9 msec
> > IW 0 [2018-08-08T13:00:18.322Z; main]: commit: done
> > IW 0 [2018-08-08T13:00:18.322Z; main]: rollback
> > IW 0 [2018-08-08T13:00:18.322Z; main]: all running merges have aborted
> > IW 0 [2018-08-08T13:00:18.323Z; main]: rollback: done finish merges
> > DW 0 [2018-08-08T13:00:18.323Z; main]: abort
> > DW 0 [2018-08-08T13:00:18.323Z; main]: done abort success=true
> > IW 0 [2018-08-08T13:00:18.323Z; main]: rollback:
> infos=_bhg(7.1.0):C108396
> > IFD 0 [2018-08-08T13:00:18.323Z; main]: now checkpoint
> > "_bhg(7.1.0):C108396" [1 segments ; isCommit = false]
> > IFD 0 [2018-08-08T13:00:18.323Z; main]: 0 msec to checkpoint
> >
> > 2018-08-07 16:38 GMT+02:00 Erick Erickson <erickerick...@gmail.com>:
> >
> >> Bjarke:
> >>
> >> One thing, what version of Solr are you moving _from_ and _to_?
> >> Solr/Lucene only guarantee one major backward revision so you can copy
> >> an index created with Solr 6 to another Solr 6 or Solr 7, but you
> >> couldn't copy an index created with Solr 5 to Solr 7...
> >>
> >> Also note that shard splitting is a very expensive operation, so be
> >> patient....
> >>
> >> Best,
> >> Erick
> >>
> >> On Tue, Aug 7, 2018 at 6:17 AM, Rahul Singh
> >> <rahul.xavier.si...@gmail.com> wrote:
> >> > Bjarke,
> >> >
> >> > I am imagining that at some point you may need to shard that data if
> it
> >> grows. Or do you imagine this data to remain stagnant?
> >> >
> >> > Generally you want to add solrcloud to do two things : 1. Increase
> >> availability with replicas 2. Increase available data via shards 3.
> >> Increase fault tolerance with leader and replicas being spread around
> the
> >> cluster.
> >> >
> >> > You would be bypassing general High availability / distributed
> computing
> >> processes by trying to not reindex.
> >> >
> >> > Rahul
> >> > On Aug 7, 2018, 7:06 AM -0400, Bjarke Buur Mortensen <
> >> morten...@eluence.com>, wrote:
> >> >> Hi List,
> >> >>
> >> >> is there a cookbook recipe for moving an existing solr core to a solr
> >> cloud
> >> >> collection.
> >> >>
> >> >> We currently have a single machine with a large core (~150gb), and we
> >> would
> >> >> like to move to solr cloud.
> >> >>
> >> >> I haven't been able to find anything that reuses an existing index,
> so
> >> any
> >> >> pointers much appreciated.
> >> >>
> >> >> Thanks,
> >> >> Bjarke
> >>
>

Reply via email to