Bjarke:

Using SPLITSHARD on an index with 6x segments just seems to not work,
even outside the standalone-> cloud issue. I'll raise a JIRA.
Meanwhile I think you'll have to re-index I'm afraid.

Thanks for raising the issue.

Erick

On Wed, Aug 8, 2018 at 6:34 AM, Bjarke Buur Mortensen
<morten...@eluence.com> wrote:
> Erick,
>
> thanks, that is of course something I left out of the original question.
> Our Solr is 7.1, so that should not present a problem (crossing fingers).
>
> However, on my dev box I'm trying out the steps, and here I have some
> segments created with version 6 of Solr.
>
> After having copied data from my non-cloud solr into my
> single-shard-single-replica collection and verified that Solr Cloud works
> with this collection, I then submit the splitshard command
>
> http://172.17.0.4:8984
> /solr/admin/collections?action=SPLITSHARD&collection=procurement&shard=shard1
>
> However, this gives me the error:
> org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException:Error
> from server at http://172.17.0.4:8984/solr:
> java.lang.IllegalArgumentException: Cannot merge a segment that has been
> created with major version 6 into this index which has been created by
> major version 7"}
>
> I have tried running both optimize and IndexUpgrader on the index before
> shard splitting, but the same error still occurs.
>
> Any ideas as to why this happens?
>
> Below is an output from running IndexUpgrader, which I cannot decipher.
> It both states that "All segments upgraded to version 7.1.0" and ''all
> running merges have aborted" ¯\_(ツ)_/¯
>
> Thanks a lot,
> Bjarke
>
>
> ======================
> java -cp
> /opt/solr/server/solr-webapp/webapp/WEB-INF/lib/lucene-backward-codecs-7.1.0.jar:/opt/solr/server/solr-webapp/webapp/WEB-INF/lib/lucene-core-7.1.0.jar
> org.apache.lucene.index.IndexUpgrader -delete-prior-commits -verbose
> /var/solr/cloud/procurement_shard1_replica_n1/data/index
> IFD 0 [2018-08-08T13:00:18.244Z; main]: init: current segments file is
> "segments_4vs";
> deletionPolicy=org.apache.lucene.index.KeepOnlyLastCommitDeletionPolicy@721e0f4f
> IFD 0 [2018-08-08T13:00:18.266Z; main]: init: load commit "segments_4vs"
> IFD 0 [2018-08-08T13:00:18.270Z; main]: now checkpoint
> "_bhg(7.1.0):C108396" [1 segments ; isCommit = false]
> IFD 0 [2018-08-08T13:00:18.270Z; main]: 0 msec to checkpoint
> IW 0 [2018-08-08T13:00:18.270Z; main]: init: create=false
> IW 0 [2018-08-08T13:00:18.273Z; main]:
> dir=MMapDirectory@/var/solr/cloud/procurement_shard1_replica_n1/data/index
> lockFactory=org.apache.lucene.store.NativeFSLockFactory@6debcae2
> index=_bhg(7.1.0):C108396
> version=7.1.0
> analyzer=null
> ramBufferSizeMB=16.0
> maxBufferedDocs=-1
> mergedSegmentWarmer=null
> delPolicy=org.apache.lucene.index.KeepOnlyLastCommitDeletionPolicy
> commit=null
> openMode=CREATE_OR_APPEND
> similarity=org.apache.lucene.search.similarities.BM25Similarity
> mergeScheduler=ConcurrentMergeScheduler: maxThreadCount=-1,
> maxMergeCount=-1, ioThrottle=true
> codec=Lucene70
> infoStream=org.apache.lucene.util.PrintStreamInfoStream
> mergePolicy=UpgradeIndexMergePolicy([TieredMergePolicy: maxMergeAtOnce=10,
> maxMergeAtOnceExplicit=30, maxMergedSegmentMB=5120.0, floorSegmentMB=2.0,
> forceMergeDeletesPctAllowed=10.0, segmentsPerTier=10.0,
> maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.1)
> indexerThreadPool=org.apache.lucene.index.DocumentsWriterPerThreadPool@5ba23b66
> readerPooling=true
> perThreadHardLimitMB=1945
> useCompoundFile=true
> commitOnClose=true
> indexSort=null
> writer=org.apache.lucene.index.IndexWriter@2ff4f00f
>
> IW 0 [2018-08-08T13:00:18.273Z; main]: MMapDirectory.UNMAP_SUPPORTED=true
> IndexUpgrader 0 [2018-08-08T13:00:18.274Z; main]: Upgrading all pre-7.1.0
> segments of index directory
> 'MMapDirectory@/var/solr/cloud/procurement_shard1_replica_n1/data/index
> lockFactory=org.apache.lucene.store.NativeFSLockFactory@6debcae2' to
> version 7.1.0...
> IW 0 [2018-08-08T13:00:18.274Z; main]: forceMerge: index now
> _bhg(7.1.0):C108396
> IW 0 [2018-08-08T13:00:18.274Z; main]: now flush at forceMerge
> IW 0 [2018-08-08T13:00:18.274Z; main]:   start flush: applyAllDeletes=true
> IW 0 [2018-08-08T13:00:18.274Z; main]:   index before flush
> _bhg(7.1.0):C108396
> DW 0 [2018-08-08T13:00:18.274Z; main]: startFullFlush
> DW 0 [2018-08-08T13:00:18.275Z; main]: main finishFullFlush success=true
> IW 0 [2018-08-08T13:00:18.275Z; main]: now apply all deletes for all
> segments buffered updates bytesUsed=0 reader pool bytesUsed=0
> BD 0 [2018-08-08T13:00:18.275Z; main]: waitApply: no deletes to apply
> UPGMP 0 [2018-08-08T13:00:18.276Z; main]: findForcedMerges:
> segmentsToUpgrade={}
> MS 0 [2018-08-08T13:00:18.282Z; main]: initDynamicDefaults spins=true
> maxThreadCount=1 maxMergeCount=6
> MS 0 [2018-08-08T13:00:18.282Z; main]: now merge
> MS 0 [2018-08-08T13:00:18.282Z; main]:   index: _bhg(7.1.0):C108396
> MS 0 [2018-08-08T13:00:18.282Z; main]:   no more merges pending; now return
> IndexUpgrader 0 [2018-08-08T13:00:18.282Z; main]: All segments upgraded to
> version 7.1.0
> IndexUpgrader 0 [2018-08-08T13:00:18.283Z; main]: Enforcing commit to
> rewrite all index metadata...
> IW 0 [2018-08-08T13:00:18.283Z; main]: commit: start
> IW 0 [2018-08-08T13:00:18.283Z; main]: commit: enter lock
> IW 0 [2018-08-08T13:00:18.283Z; main]: commit: now prepare
> IW 0 [2018-08-08T13:00:18.283Z; main]: prepareCommit: flush
> IW 0 [2018-08-08T13:00:18.283Z; main]:   index before flush
> _bhg(7.1.0):C108396
> DW 0 [2018-08-08T13:00:18.283Z; main]: startFullFlush
> IW 0 [2018-08-08T13:00:18.283Z; main]: now apply all deletes for all
> segments buffered updates bytesUsed=0 reader pool bytesUsed=0
> BD 0 [2018-08-08T13:00:18.283Z; main]: waitApply: no deletes to apply
> DW 0 [2018-08-08T13:00:18.284Z; main]: main finishFullFlush success=true
> IW 0 [2018-08-08T13:00:18.284Z; main]: startCommit(): start
> IW 0 [2018-08-08T13:00:18.284Z; main]: startCommit
> index=_bhg(7.1.0):C108396 changeCount=2
> IW 0 [2018-08-08T13:00:18.293Z; main]: startCommit: wrote pending segments
> file "pending_segments_4vt"
> IW 0 [2018-08-08T13:00:18.295Z; main]: done all syncs:
> [_bhg_Lucene50_0.tip, _bhg.fdx, _bhg.fnm, _bhg.nvm, _bhg.fdt, _bhg.si,
> _bhg_Lucene50_0.pos, _bhg.nvd, _bhg_Lucene50_0.doc, _bhg_Lucene50_0.tim]
> IW 0 [2018-08-08T13:00:18.295Z; main]: commit: pendingCommit != null
> IW 0 [2018-08-08T13:00:18.298Z; main]: commit: done writing segments file
> "segments_4vt"
> IFD 0 [2018-08-08T13:00:18.298Z; main]: now checkpoint
> "_bhg(7.1.0):C108396" [1 segments ; isCommit = true]
> IFD 0 [2018-08-08T13:00:18.298Z; main]: deleteCommits: now decRef commit
> "segments_4vs"
> IFD 0 [2018-08-08T13:00:18.298Z; main]: delete [segments_4vs]
> IFD 0 [2018-08-08T13:00:18.299Z; main]: 0 msec to checkpoint
> IW 0 [2018-08-08T13:00:18.319Z; main]: commit: took 16.0 msec
> IW 0 [2018-08-08T13:00:18.319Z; main]: commit: done
> IndexUpgrader 0 [2018-08-08T13:00:18.319Z; main]: Committed upgraded
> metadata to index.
> IW 0 [2018-08-08T13:00:18.319Z; main]: now flush at close
> IW 0 [2018-08-08T13:00:18.319Z; main]:   start flush: applyAllDeletes=true
> IW 0 [2018-08-08T13:00:18.319Z; main]:   index before flush
> _bhg(7.1.0):C108396
> DW 0 [2018-08-08T13:00:18.319Z; main]: startFullFlush
> DW 0 [2018-08-08T13:00:18.320Z; main]: main finishFullFlush success=true
> IW 0 [2018-08-08T13:00:18.320Z; main]: now apply all deletes for all
> segments buffered updates bytesUsed=0 reader pool bytesUsed=0
> BD 0 [2018-08-08T13:00:18.320Z; main]: waitApply: no deletes to apply
> MS 0 [2018-08-08T13:00:18.320Z; main]: updateMergeThreads ioThrottle=true
> targetMBPerSec=10240.0 MB/sec
> MS 0 [2018-08-08T13:00:18.320Z; main]: now merge
> MS 0 [2018-08-08T13:00:18.321Z; main]:   index: _bhg(7.1.0):C108396
> MS 0 [2018-08-08T13:00:18.321Z; main]:   no more merges pending; now return
> IW 0 [2018-08-08T13:00:18.321Z; main]: waitForMerges
> IW 0 [2018-08-08T13:00:18.321Z; main]: waitForMerges done
> IW 0 [2018-08-08T13:00:18.321Z; main]: commit: start
> IW 0 [2018-08-08T13:00:18.321Z; main]: commit: enter lock
> IW 0 [2018-08-08T13:00:18.321Z; main]: commit: now prepare
> IW 0 [2018-08-08T13:00:18.321Z; main]: prepareCommit: flush
> IW 0 [2018-08-08T13:00:18.321Z; main]:   index before flush
> _bhg(7.1.0):C108396
> DW 0 [2018-08-08T13:00:18.321Z; main]: startFullFlush
> IW 0 [2018-08-08T13:00:18.321Z; main]: now apply all deletes for all
> segments buffered updates bytesUsed=0 reader pool bytesUsed=0
> BD 0 [2018-08-08T13:00:18.322Z; main]: waitApply: no deletes to apply
> DW 0 [2018-08-08T13:00:18.322Z; main]: main finishFullFlush success=true
> IW 0 [2018-08-08T13:00:18.322Z; main]: startCommit(): start
> IW 0 [2018-08-08T13:00:18.322Z; main]:   skip startCommit(): no changes
> pending
> IW 0 [2018-08-08T13:00:18.322Z; main]: commit: pendingCommit == null; skip
> IW 0 [2018-08-08T13:00:18.322Z; main]: commit: took 0.9 msec
> IW 0 [2018-08-08T13:00:18.322Z; main]: commit: done
> IW 0 [2018-08-08T13:00:18.322Z; main]: rollback
> IW 0 [2018-08-08T13:00:18.322Z; main]: all running merges have aborted
> IW 0 [2018-08-08T13:00:18.323Z; main]: rollback: done finish merges
> DW 0 [2018-08-08T13:00:18.323Z; main]: abort
> DW 0 [2018-08-08T13:00:18.323Z; main]: done abort success=true
> IW 0 [2018-08-08T13:00:18.323Z; main]: rollback: infos=_bhg(7.1.0):C108396
> IFD 0 [2018-08-08T13:00:18.323Z; main]: now checkpoint
> "_bhg(7.1.0):C108396" [1 segments ; isCommit = false]
> IFD 0 [2018-08-08T13:00:18.323Z; main]: 0 msec to checkpoint
>
> 2018-08-07 16:38 GMT+02:00 Erick Erickson <erickerick...@gmail.com>:
>
>> Bjarke:
>>
>> One thing, what version of Solr are you moving _from_ and _to_?
>> Solr/Lucene only guarantee one major backward revision so you can copy
>> an index created with Solr 6 to another Solr 6 or Solr 7, but you
>> couldn't copy an index created with Solr 5 to Solr 7...
>>
>> Also note that shard splitting is a very expensive operation, so be
>> patient....
>>
>> Best,
>> Erick
>>
>> On Tue, Aug 7, 2018 at 6:17 AM, Rahul Singh
>> <rahul.xavier.si...@gmail.com> wrote:
>> > Bjarke,
>> >
>> > I am imagining that at some point you may need to shard that data if it
>> grows. Or do you imagine this data to remain stagnant?
>> >
>> > Generally you want to add solrcloud to do two things : 1. Increase
>> availability with replicas 2. Increase available data via shards 3.
>> Increase fault tolerance with leader and replicas being spread around the
>> cluster.
>> >
>> > You would be bypassing general High availability / distributed computing
>> processes by trying to not reindex.
>> >
>> > Rahul
>> > On Aug 7, 2018, 7:06 AM -0400, Bjarke Buur Mortensen <
>> morten...@eluence.com>, wrote:
>> >> Hi List,
>> >>
>> >> is there a cookbook recipe for moving an existing solr core to a solr
>> cloud
>> >> collection.
>> >>
>> >> We currently have a single machine with a large core (~150gb), and we
>> would
>> >> like to move to solr cloud.
>> >>
>> >> I haven't been able to find anything that reuses an existing index, so
>> any
>> >> pointers much appreciated.
>> >>
>> >> Thanks,
>> >> Bjarke
>>

Reply via email to