To be more concrete: Is the definitive test of whether or not a core's
index is corrupt to copy it onto a new set of hardware and attempt to write
to it? If this is a definitive test, we can run the experiment and update
the report so you have a sense of how often this happens.

Since this is a SOLR cloud node, which is already removed but whose data
dir was preserved, I believe I can just copy the data directory to a fresh
machine and start a regular non-cloud solr node hosting this core. Can you
please confirm that this will be a definitive test, or whether there is
some aspect needed to make it definitive?

Thanks!

On Wed, Oct 3, 2018 at 2:10 AM Stephen Bianamara <sbianam...@panopto.com>
wrote:

> Hello All --
>
> As it would happen, we've seen this error on version 6.6.2 very recently.
> This is also on an AWS instance, like Simon's report. The drive doesn't
> show any sign of being unhealthy, either from cursory investigation. FWIW,
> this occurred during a collection backup.
>
> Erick, is there some diagnostic data we can find to help pin this down?
>
> Thanks!
> Stephen
>
> On Sun, Sep 30, 2018 at 12:48 PM Susheel Kumar <susheel2...@gmail.com>
> wrote:
>
>> Thank you, Simon. Which basically points that something related to env and
>> was causing the checksum failures than any lucene/solr issue.
>>
>> Eric - I did check with hardware folks and they are aware of some VMware
>> issue where the VM hosted in HCI environment is coming into some halt
>> state
>> for minute or so and may be loosing connections to disk/network.  So that
>> probably may be the reason of index corruption though they have not been
>> able to find anything specific from logs during the time Solr run into
>> issue
>>
>> Also I had again issue where Solr is loosing the connection with zookeeper
>> (Client session timed out, have not heard from server in 8367ms for
>> sessionid 0x0)  Does that points to similar hardware issue, Any
>> suggestions?
>>
>> Thanks,
>> Susheel
>>
>> 2018-09-29 17:30:44.070 INFO
>> (searcherExecutor-7-thread-1-processing-n:server54:8080_solr
>> x:COLL_shard4_replica2 s:shard4 c:COLL r:core_node8) [c:COLL s:shard4
>> r:core_node8 x:COLL_shard4_replica2] o.a.s.c.SolrCore
>> [COLL_shard4_replica2] Registered new searcher
>> Searcher@7a4465b1[COLL_shard4_replica2]
>>
>> main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_7x3f(6.6.2):C826923/317917:delGen=2523)
>> Uninverting(_83pb(6.6.2):C805451/172968:delGen=2957)
>> Uninverting(_3ywj(6.6.2):C727978/334529:delGen=2962)
>> Uninverting(_7vsw(6.6.2):C872110/385178:delGen=2020)
>> Uninverting(_8n89(6.6.2):C741293/109260:delGen=3863)
>> Uninverting(_7zkq(6.6.2):C720666/101205:delGen=3151)
>> Uninverting(_825d(6.6.2):C707731/112410:delGen=3168)
>> Uninverting(_dgwu(6.6.2):C760421/295964:delGen=4624)
>> Uninverting(_gs5x(6.6.2):C540942/138952:delGen=1623)
>> Uninverting(_gu6a(6.6.2):c75213/35640:delGen=1110)
>> Uninverting(_h33i(6.6.2):c131276/40356:delGen=706)
>> Uninverting(_h5tc(6.6.2):c44320/11080:delGen=380)
>> Uninverting(_h9d9(6.6.2):c35088/3188:delGen=104)
>> Uninverting(_h80h(6.6.2):c11927/3412:delGen=153)
>> Uninverting(_h7ll(6.6.2):c11284/1368:delGen=205)
>> Uninverting(_h8bs(6.6.2):c11518/2103:delGen=149)
>> Uninverting(_h9r3(6.6.2):c16439/1018:delGen=52)
>> Uninverting(_h9z1(6.6.2):c9428/823:delGen=27)
>> Uninverting(_h9v2(6.6.2):c933/33:delGen=12)
>> Uninverting(_ha1c(6.6.2):c1056/1:delGen=1)
>> Uninverting(_ha6i(6.6.2):c1883/124:delGen=8)
>> Uninverting(_ha3x(6.6.2):c807/14:delGen=3)
>> Uninverting(_ha47(6.6.2):c1229/133:delGen=6)
>> Uninverting(_hapk(6.6.2):c523) Uninverting(_haoq(6.6.2):c279)
>> Uninverting(_hamr(6.6.2):c311) Uninverting(_hap0(6.6.2):c338)
>> Uninverting(_hapu(6.6.2):c275) Uninverting(_hapv(6.6.2):C4/2:delGen=1)
>> Uninverting(_hapw(6.6.2):C5/2:delGen=1)
>> Uninverting(_hapx(6.6.2):C2/1:delGen=1)
>> Uninverting(_hapy(6.6.2):C2/1:delGen=1)
>> Uninverting(_hapz(6.6.2):C3/1:delGen=1)
>> Uninverting(_haq0(6.6.2):C6/3:delGen=1)
>> Uninverting(_haq1(6.6.2):C1)))}
>> 2018-09-29 17:30:52.390 WARN
>>
>> (zkCallback-5-thread-91-processing-n:server54:8080_solr-SendThread(server117:2182))
>> [   ] o.a.z.ClientCnxn Client session timed out, have not heard from
>> server in 8367ms for sessionid 0x0
>> 2018-09-29 17:31:01.302 WARN
>>
>> (zkCallback-5-thread-91-processing-n:server54:8080_solr-SendThread(server120:2182))
>> [   ] o.a.z.ClientCnxn Client session timed out, have not heard from
>> server in 8812ms for sessionid 0x0
>> 2018-09-29 17:31:14.049 INFO
>> (zkCallback-5-thread-91-processing-n:server54:8080_solr-EventThread) [
>>   ] o.a.s.c.c.ConnectionManager Connection with ZooKeeper
>> reestablished.
>> 2018-09-29 17:31:14.049 INFO
>> (zkCallback-5-thread-91-processing-n:server54:8080_solr-EventThread) [
>>   ] o.a.s.c.ZkController ZooKeeper session re-connected ... refreshing
>> core states after session expiration.
>> 2018-09-29 17:31:14.051 INFO
>> (zkCallback-5-thread-91-processing-n:server54:8080_solr-EventThread) [
>>   ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (16)
>> -> (15)
>> 2018-09-29 17:31:14.144 INFO  (qtp834133664-520378) [c:COLL s:shard4
>> r:core_node8 x:COLL_shard4_replica2] o.a.s.c.S.Request
>> [COLL_shard4_replica2]  webapp=/solr path=/admin/ping
>>
>> params={distrib=false&df=wordTokens&_stateVer_=COLL:1246&preferLocalShards=false&qt=/admin/ping&fl=id&fl=score&shards.purpose=4&start=0&fsv=true&shard.url=
>> http://server54:8080/solr/COLL_shard4_replica2/|http://server53:8080/solr/COLL_shard4_replica1/&rows=10&version=2&q={!lucene}*:*&NOW=1538242274139&isShard=true&wt=javabin
>> }
>> webapp=/solr path=/admin/ping
>>
>> params={distrib=false&df=wordTokens&_stateVer_=COLL:1246&preferLocalShards=false&qt=/admin/ping&fl=id&fl=score&shards.purpose=4&start=0&fsv=true&shard.url=
>> http://server54:8080/solr/COLL_shard4_replica2/|http://server53:8080/solr/COLL_shard4_replica1/&rows=10&version=2&q={!lucene}*:*&NOW=1538242274139&isShard=true&wt=javabin
>> }
>> hits=4989979 status=0 QTime=0
>>
>>
>>
>>
>> On Wed, Sep 26, 2018 at 9:44 AM simon <mtnes...@gmail.com> wrote:
>>
>> > I saw something like this a year ago which i reported as a possible
>> bug  (
>> > https://issues.apache.org/jira/browse/SOLR-10840, which has  a full
>> > description and stack traces)
>> >
>> > This occurred very randomly on an AWS instance; moving the index
>> directory
>> > to a different file system did not fix the problem Eventually I cloned
>> our
>> > environment to a new AWS instance, which proved to be the solution.
>> Why, I
>> > have no idea...
>> >
>> > -Simon
>> >
>> > On Mon, Sep 24, 2018 at 1:13 PM, Susheel Kumar <susheel2...@gmail.com>
>> > wrote:
>> >
>> > > Got it. I'll have first hardware folks check and if they don't
>> see/find
>> > > anything suspicious then i'll return here.
>> > >
>> > > Wondering if any body has seen similar error and if they were able to
>> > > confirm if it was hardware fault or so.
>> > >
>> > > Thnx
>> > >
>> > > On Mon, Sep 24, 2018 at 1:01 PM Erick Erickson <
>> erickerick...@gmail.com>
>> > > wrote:
>> > >
>> > > > Mind you it could _still_ be Solr/Lucene, but let's check the
>> hardware
>> > > > first ;)
>> > > > On Mon, Sep 24, 2018 at 9:50 AM Susheel Kumar <
>> susheel2...@gmail.com>
>> > > > wrote:
>> > > > >
>> > > > > Hi Erick,
>> > > > >
>> > > > > Thanks so much for your reply.  I'll now look mostly into any
>> > possible
>> > > > > hardware issues than Solr/Lucene.
>> > > > >
>> > > > > Thanks again.
>> > > > >
>> > > > > On Mon, Sep 24, 2018 at 12:43 PM Erick Erickson <
>> > > erickerick...@gmail.com
>> > > > >
>> > > > > wrote:
>> > > > >
>> > > > > > There are several of reasons this would "suddenly" start
>> appearing.
>> > > > > > 1> Your disk went bad and some sector is no longer faithfully
>> > > > > > recording the bits. In this case the checksum will be wrong
>> > > > > > 2> You ran out of disk space sometime and the index was
>> corrupted.
>> > > > > > This isn't really a hardware problem.
>> > > > > > 3> Your disk controller is going wonky and not reading reliably.
>> > > > > >
>> > > > > > The "possible hardware issue" message is to alert you that this
>> is
>> > > > > > highly unusual and you should at leasts consider doing integrity
>> > > > > > checks on your disk before assuming it's a Solr/Lucene problem
>> > > > > >
>> > > > > > Best,
>> > > > > > Erick
>> > > > > > On Mon, Sep 24, 2018 at 9:26 AM Susheel Kumar <
>> > susheel2...@gmail.com
>> > > >
>> > > > > > wrote:
>> > > > > > >
>> > > > > > > Hello,
>> > > > > > >
>> > > > > > > I am still trying to understand the corrupt index exception we
>> > saw
>> > > > in our
>> > > > > > > logs. What does the hardware problem comment indicates here?
>> > Does
>> > > > that
>> > > > > > > mean it caused most likely due to hardware issue?
>> > > > > > >
>> > > > > > > We never had this problem in last couple of months. The Solr
>> is
>> > > > 6.6.2 and
>> > > > > > > ZK: 3.4.10.
>> > > > > > >
>> > > > > > > Please share your thoughts.
>> > > > > > >
>> > > > > > > Thanks,
>> > > > > > > Susheel
>> > > > > > >
>> > > > > > > Caused by: org.apache.lucene.index.CorruptIndexException:
>> > checksum
>> > > > > > > failed *(hardware
>> > > > > > > problem?)* : expected=db243d1a actual=7a00d3d2
>> > > > > > >
>> > > > > >
>> > > > (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/
>> > > app/solr/data/COLL_shard1_replica1/data/index/_i27s.cfs")
>> > > > > > > [slice=_i27s_Lucene50_0.tim])
>> > > > > > >
>> > > > > > > It suddenly started in the logs and before which there was no
>> > such
>> > > > error.
>> > > > > > > Searches & ingestions all seems to be working prior to that.
>> > > > > > >
>> > > > > > > ----
>> > > > > > >
>> > > > > > > 2018-09-03 17:16:49.056 INFO  (qtp834133664-519872) [c:COLL
>> > > s:shard1
>> > > > > > > r:core_node1 x:COLL_shard1_replica1]
>> > > > > > > o.a.s.u.p.StatelessScriptUpdateProcessorFactory
>> > > > update-script#processAdd:
>> > > > > > >
>> > > > newid=G31MXMRZESC0CYPR!A-G31MXMRZESC0CYPR.2552019802_1-25520
>> > > 08480_1-en_US
>> > > > > > > 2018-09-03 17:16:49.057 ERROR (qtp834133664-519872) [c:COLL
>> > > s:shard1
>> > > > > > > r:core_node1 x:COLL_shard1_replica1]
>> o.a.s.h.RequestHandlerBase
>> > > > > > > org.apache.solr.common.SolrException: Exception writing
>> document
>> > > id
>> > > > > > > G31MXMRZESC0CYPR!A-G31MXMRZESC0CYPR.2552019802_1-2552008480_
>> > > 1-en_US
>> > > > to
>> > > > > > the
>> > > > > > > index; possible analysis error.
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpd
>> > > ateHandler2.java:206)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.processor.RunUpdateProcessor.processA
>> > > dd(RunUpdateProcessorFactory.java:67)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.processor.UpdateRequestProcessor.proc
>> > > essAdd(UpdateRequestProcessor.java:55)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.processor.DistributedUpdateProcessor.
>> > > doLocalAdd(DistributedUpdateProcessor.java:979)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.processor.DistributedUpdateProcessor.
>> > > versionAdd(DistributedUpdateProcessor.java:1192)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.processor.DistributedUpdateProcessor.
>> > > processAdd(DistributedUpdateProcessor.java:748)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.processor.UpdateRequestProcessor.proc
>> > > essAdd(UpdateRequestProcessor.java:55)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > >
>> org.apache.solr.update.processor.StatelessScriptUpdateProcessorFactory$
>> > > ScriptUpdateProcessor.processAdd(StatelessScriptUpdateProces
>> > > sorFactory.java:380)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.handler.loader.JavabinLoader$1.update(Javabi
>> > > nLoader.java:98)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.client.solrj.request.JavaBinUpdateRequestCod
>> > > ec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:180)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.client.solrj.request.JavaBinUpdateRequestCod
>> > > ec$1.readIterator(JavaBinUpdateRequestCodec.java:136)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.common.util.JavaBinCodec.readObject(JavaBinC
>> > > odec.java:306)
>> > > > > > > at
>> > > > > > org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCode
>> > > c.java:251)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.client.solrj.request.JavaBinUpdateRequestCod
>> > > ec$1.readNamedList(JavaBinUpdateRequestCodec.java:122)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.common.util.JavaBinCodec.readObject(JavaBinC
>> > > odec.java:271)
>> > > > > > > at
>> > > > > > org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCode
>> > > c.java:251)
>> > > > > > > at
>> > > > > >
>> > > > org.apache.solr.common.util.JavaBinCodec.unmarshal(JavaBinCo
>> > > dec.java:173)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.client.solrj.request.JavaBinUpdateRequestCod
>> > > ec.unmarshal(JavaBinUpdateRequestCodec.java:187)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.handler.loader.JavabinLoader.parseAndLoadDoc
>> > > s(JavabinLoader.java:108)
>> > > > > > > at
>> > > > > >
>> > > >
>> > org.apache.solr.handler.loader.JavabinLoader.load(JavabinLoader.java:55)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRe
>> > > questHandler.java:97)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.handler.ContentStreamHandlerBase.handleReque
>> > > stBody(ContentStreamHandlerBase.java:68)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.handler.RequestHandlerBase.handleRequest(Req
>> > > uestHandlerBase.java:173)
>> > > > > > > at org.apache.solr.core.SolrCore.execute(SolrCore.java:2477)
>> > > > > > > at
>> > > > org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:723)
>> > > > > > > at
>> org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:
>> > > 529)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDisp
>> > > atchFilter.java:361)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDisp
>> > > atchFilter.java:305)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilte
>> > > r(ServletHandler.java:1691)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHan
>> > > dler.java:582)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.handler.ScopedHandler.handle(Scoped
>> > > Handler.java:143)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.security.SecurityHandler.handle(SecurityHa
>> > > ndler.java:548)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.session.SessionHandler.doHandle(
>> > > SessionHandler.java:226)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.handler.ContextHandler.doHandle(
>> > > ContextHandler.java:1180)
>> > > > > > > at
>> > > > > >
>> > > > org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHand
>> > > ler.java:512)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.session.SessionHandler.doScope(
>> > > SessionHandler.java:185)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.handler.ContextHandler.doScope(
>> > > ContextHandler.java:1112)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.handler.ScopedHandler.handle(Scoped
>> > > Handler.java:141)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.handler.ContextHandlerCollection.ha
>> > > ndle(ContextHandlerCollection.java:213)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.handler.HandlerCollection.handle(
>> > > HandlerCollection.java:119)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.handler.HandlerWrapper.handle(Handl
>> > > erWrapper.java:134)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.rewrite.handler.RewriteHandler.handle(Rewr
>> > > iteHandler.java:335)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.handler.HandlerWrapper.handle(Handl
>> > > erWrapper.java:134)
>> > > > > > > at org.eclipse.jetty.server.Server.handle(Server.java:534)
>> > > > > > > at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.
>> > > java:320)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.server.HttpConnection.onFillable(HttpConne
>> > > ction.java:251)
>> > > > > > > at
>> > > > > > > org.eclipse.jetty.io
>> > > > > > .AbstractConnection$ReadCallback.succeeded(AbstractConnectio
>> > > n.java:273)
>> > > > > > > at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.
>> > > java:95)
>> > > > > > > at
>> > > > > > > org.eclipse.jetty.io
>> > > > > > .SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume
>> > > .executeProduceConsume(ExecuteProduceConsume.java:303)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume
>> > > .produceConsume(ExecuteProduceConsume.java:148)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume
>> > > .run(ExecuteProduceConsume.java:136)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(Queued
>> > > ThreadPool.java:671)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedT
>> > > hreadPool.java:589)
>> > > > > > > at java.lang.Thread.run(Thread.java:748)
>> > > > > > > Caused by: org.apache.lucene.store.AlreadyClosedException:
>> this
>> > > > > > IndexWriter
>> > > > > > > is closed
>> > > > > > > at
>> > > > org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:749)
>> > > > > > > at
>> > > > org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:763)
>> > > > > > > at
>> > > > > >
>> > > > org.apache.lucene.index.IndexWriter.updateDocument(IndexWrit
>> > > er.java:1567)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.DirectUpdateHandler2.updateDocument(D
>> > > irectUpdateHandler2.java:924)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.DirectUpdateHandler2.updateDocOrDocVa
>> > > lues(DirectUpdateHandler2.java:913)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.DirectUpdateHandler2.doNormalUpdate(D
>> > > irectUpdateHandler2.java:302)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.DirectUpdateHandler2.addDoc0(DirectUp
>> > > dateHandler2.java:239)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpd
>> > > ateHandler2.java:194)
>> > > > > > > ... 54 more
>> > > > > > > Caused by: org.apache.lucene.index.CorruptIndexException:
>> > checksum
>> > > > failed
>> > > > > > > (hardware problem?) : expected=db243d1a actual=7a00d3d2
>> > > > > > >
>> > > > > >
>> > > > (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/
>> > > app/solr/data/COLL_shard1_replica1/data/index/_i27s.cfs")
>> > > > > > > [slice=_i27s_Lucene50_0.tim]))
>> > > > > > > at org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.
>> > > java:419)
>> > > > > > > at
>> > > > > >
>> > > > org.apache.lucene.codecs.CodecUtil.checksumEntireFile(CodecU
>> > > til.java:526)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.lucene.codecs.blocktree.BlockTreeTermsReader.chec
>> > > kIntegrity(BlockTreeTermsReader.java:336)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$Fie
>> > > ldsReader.checkIntegrity(PerFieldPostingsFormat.java:348)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.lucene.codecs.perfield.PerFieldMergeState$FilterF
>> > > ieldsProducer.checkIntegrity(PerFieldMergeState.java:271)
>> > > > > > > at
>> > > >
>> org.apache.lucene.codecs.FieldsConsumer.merge(FieldsConsumer.java:96)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$Fie
>> > > ldsWriter.merge(PerFieldPostingsFormat.java:164)
>> > > > > > > at
>> > > > > >
>> > > >
>> > org.apache.lucene.index.SegmentMerger.mergeTerms(SegmentMerger.java:216)
>> > > > > > > at
>> > > > org.apache.lucene.index.SegmentMerger.merge(SegmentMerger.java:101)
>> > > > > > > at
>> > > >
>> org.apache.lucene.index.IndexWriter.mergeMiddle(IndexWriter.java:4356)
>> > > > > > > at org.apache.lucene.index.IndexWriter.merge(IndexWriter.java:
>> > > 3931)
>> > > > > > > at
>> > > >
>> org.apache.solr.update.SolrIndexWriter.merge(SolrIndexWriter.java:188)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.lucene.index.ConcurrentMergeScheduler.doMerge(Con
>> > > currentMergeScheduler.java:624)
>> > > > > > > at
>> > > > > > >
>> > > > > >
>> > > > org.apache.lucene.index.ConcurrentMergeScheduler$MergeThread
>> > > .run(ConcurrentMergeScheduler.java:661)
>> > > > > > >
>> > > > > > > 2018-09-03 17:16:49.116 INFO  (qtp834133664-519872) [c:COLL
>> > > s:shard1
>> > > > > > > r:core_node1 x:COLL_shard1_replica1] o.a.s.c.S.Request
>> > > > > > > [COLL_shard1_replica1]  webapp=/solr path=/update
>> > > > > > > params={wt=javabin&version=2} status=400 QTime=69
>> > > > > >
>> > > >
>> > >
>> >
>>
>

Reply via email to