and there is document id mentioned above when it failed with analysis error. You can look how those documents differ as Eric suggested.
On Mon, Jul 17, 2017 at 11:53 AM, Erick Erickson <erickerick...@gmail.com> wrote: > Joe: > > I agree that 46 million docs later you'd expect things to have settled > out. However, I do note that you have > "add-unknown-fields-to-the-schema" in your error stack which means > you're using "field guessing", sometimes called data_driven. I would > recommend you do _not_ use this for production as, while it does the > best job it can it has to make assumptions about what the data looks > like based on the first document it sees which may later be violated. > Getting "possible analysis error" is one of the messages that happens > when this occurs. > > The simple example is that if the first time data_driven sees "1" > it'll guess integer. If sometime later there's a doc with "1.0" it'll > generate a parse error. > > I totally agree that 46 million docs later you'd expect all of this > kind of thing to have flushed out, but the "possible analysis error" > seems to be pointing that direction. If this is, indeed, the problem > you'll see better evidence on the Solr instance that's actually having > the problem. Unfortunately you'll just to look at one Solr log from > each shard to see whether this is an issue. > > Best, > Erick > > On Mon, Jul 17, 2017 at 7:23 AM, Joe Obernberger > <joseph.obernber...@gmail.com> wrote: > > So far we've indexed about 46 million documents, but over the weekend, > these > > errors started coming up. I would expect that if there was a basic > issue, > > it would have started right away? We ran a test cluster with just a few > > shards/replicas prior and didn't see any issues using the same indexing > > code, but we're running a lot more indexers simultaneously with the > larger > > cluster; perhaps we're just overloading HDFS? The same nodes that run > Solr > > also run HDFS datanodes, but they are pretty beefy machines; we're not > > swapping. > > > > As Shawn pointed out, I will be checking the HDFS version (we're using > > Cloudera CDH 5.10.2), and the HDFS logs. > > > > -Joe > > > > > > > > On 7/17/2017 10:16 AM, Susheel Kumar wrote: > >> > >> There is some analysis error also. I would suggest to test the indexer > on > >> just one shard setup first, then test for a replica (1 shard and 1 > >> replica) > >> and then test for 2 shards and 2 replica. This would confirm if there > is > >> basic issue with indexing / cluster setup. > >> > >> On Mon, Jul 17, 2017 at 9:04 AM, Joe Obernberger < > >> joseph.obernber...@gmail.com> wrote: > >> > >>> Some more info: > >>> > >>> When I stop all the indexers, in about 5-10 minutes the cluster goes > all > >>> green. When I start just one indexer, several nodes immediately go > down > >>> with the 'Error adding log' message. > >>> > >>> I'm using CloudSolrClient.add(List<SolrInputDocument>) to do the > >>> indexing. Is this correct for SolrCloud? > >>> > >>> Thank you! > >>> > >>> -Joe > >>> > >>> > >>> > >>> On 7/17/2017 8:36 AM, Joe Obernberger wrote: > >>> > >>>> We've been indexing data on a 45 node cluster with 100 shards and 3 > >>>> replicas, but our indexing processes have been stopping due to errors. > >>>> On > >>>> the server side the error is "Error logging add". Stack trace: > >>>> > >>>> 2017-07-17 12:29:24.057 INFO (qtp985934102-5161548) [c:UNCLASS > >>>> s:shard58 > >>>> r:core_node290 x:UNCLASS_shard58_replica1] > >>>> o.a.s.u.p.LogUpdateProcessorFactory > >>>> [UNCLASS_shard58_replica1] webapp=/solr path=/update > >>>> params={update.distrib=FROMLEADER&update.chain=add-unknown- > >>>> fields-to-the-schema&distrib.from=http://tarvos:9100/solr/ > >>>> UNCLASS_shard58_replica2/&wt=javabin&version=2}{add=[ > >>>> COLLECT20003218348784 (1573172872544780288), COLLECT20003218351447 > >>>> (1573172872620277760), COLLECT20003218353085 (1573172872625520640), > >>>> COLLECT20003218357937 (1573172872627617792), COLLECT20003218361860 > >>>> (1573172872629714944), COLLECT20003218362535 (1573172872631812096)]} 0 > >>>> 171 > >>>> 2017-07-17 12:29:24.160 INFO (qtp985934102-5160762) [c:UNCLASS > >>>> s:shard13 > >>>> r:core_node81 x:UNCLASS_shard13_replica1] > >>>> o.a.s.u.p.LogUpdateProcessorFactory > >>>> [UNCLASS_shard13_replica1] webapp=/solr path=/update > >>>> params={update.distrib=FROMLEADER&update.chain=add-unknown- > >>>> fields-to-the-schema&distrib.from=http://tarvos:9100/solr/ > >>>> UNCLASS_shard13_replica2/&wt=javabin&version=2}{add=[ > >>>> COLLECT20003218344436 (1573172872538488832), COLLECT20003218347497 > >>>> (1573172872620277760), COLLECT20003218351645 (1573172872625520640), > >>>> COLLECT20003218356965 (1573172872629714944), COLLECT20003218357775 > >>>> (1573172872632860672), COLLECT20003218358017 (1573172872646492160), > >>>> COLLECT20003218358152 (1573172872650686464), COLLECT20003218359395 > >>>> (1573172872651735040), COLLECT20003218362571 (1573172872652783616)]} 0 > >>>> 274 > >>>> 2017-07-17 12:29:24.163 INFO (qtp985934102-5161057) [c:UNCLASS > >>>> s:shard43 > >>>> r:core_node108 x:UNCLASS_shard43_replica1] > >>>> o.a.s.u.p.LogUpdateProcessorFactory > >>>> [UNCLASS_shard43_replica1] webapp=/solr path=/update > >>>> params={update.distrib=FROMLEADER&update.chain=add-unknown- > >>>> fields-to-the-schema&distrib.from=http://tarvos:9100/solr/ > >>>> UNCLASS_shard43_replica2/&wt=javabin&version=2}{} 0 0 > >>>> 2017-07-17 12:29:24.164 ERROR (qtp985934102-5161057) [c:UNCLASS > >>>> s:shard43 > >>>> r:core_node108 x:UNCLASS_shard43_replica1] o.a.s.h.RequestHandlerBase > >>>> org.apache.solr.common.SolrException: Error logging add > >>>> at org.apache.solr.update.TransactionLog.write( > TransactionLog. > >>>> java:418) > >>>> at org.apache.solr.update.UpdateLog.add(UpdateLog.java:532) > >>>> at org.apache.solr.update.UpdateLog.add(UpdateLog.java:516) > >>>> at org.apache.solr.update.processor. > DistributedUpdateProcessor. > >>>> versionAdd(DistributedUpdateProcessor.java:1113) > >>>> at org.apache.solr.update.processor. > DistributedUpdateProcessor. > >>>> processAdd(DistributedUpdateProcessor.java:748) > >>>> at org.apache.solr.update.processor. > LogUpdateProcessorFactory$L > >>>> ogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:103) > >>>> at org.apache.solr.handler.loader.JavabinLoader$1.update( > Javabi > >>>> nLoader.java:98) > >>>> at org.apache.solr.client.solrj.request. > JavaBinUpdateRequestCod > >>>> ec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:180) > >>>> at org.apache.solr.client.solrj.request. > JavaBinUpdateRequestCod > >>>> ec$1.readIterator(JavaBinUpdateRequestCodec.java:136) > >>>> at org.apache.solr.common.util.JavaBinCodec.readObject( > JavaBinC > >>>> odec.java:306) > >>>> at org.apache.solr.common.util.JavaBinCodec.readVal( > JavaBinCode > >>>> c.java:251) > >>>> at org.apache.solr.client.solrj.request. > JavaBinUpdateRequestCod > >>>> ec$1.readNamedList(JavaBinUpdateRequestCodec.java:122) > >>>> at org.apache.solr.common.util.JavaBinCodec.readObject( > JavaBinC > >>>> odec.java:271) > >>>> at org.apache.solr.common.util.JavaBinCodec.readVal( > JavaBinCode > >>>> c.java:251) > >>>> at org.apache.solr.common.util.JavaBinCodec.unmarshal( > JavaBinCo > >>>> dec.java:173) > >>>> at org.apache.solr.client.solrj.request. > JavaBinUpdateRequestCod > >>>> ec.unmarshal(JavaBinUpdateRequestCodec.java:187) > >>>> at org.apache.solr.handler.loader.JavabinLoader. > parseAndLoadDoc > >>>> s(JavabinLoader.java:108) > >>>> at org.apache.solr.handler.loader.JavabinLoader.load( > JavabinLoa > >>>> der.java:55) > >>>> at org.apache.solr.handler.UpdateRequestHandler$1.load( > UpdateRe > >>>> questHandler.java:97) > >>>> at org.apache.solr.handler.ContentStreamHandlerBase. > handleReque > >>>> stBody(ContentStreamHandlerBase.java:68) > >>>> at org.apache.solr.handler.RequestHandlerBase. > handleRequest(Req > >>>> uestHandlerBase.java:173) > >>>> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2477) > >>>> at org.apache.solr.servlet.HttpSolrCall.execute( > HttpSolrCall. > >>>> java:723) > >>>> at org.apache.solr.servlet.HttpSolrCall.call( > HttpSolrCall.java: > >>>> 529) > >>>> at org.apache.solr.servlet.SolrDispatchFilter.doFilter( > SolrDisp > >>>> atchFilter.java:361) > >>>> at org.apache.solr.servlet.SolrDispatchFilter.doFilter( > SolrDisp > >>>> atchFilter.java:305) > >>>> at org.eclipse.jetty.servlet.ServletHandler$CachedChain. > doFilte > >>>> r(ServletHandler.java:1691) > >>>> at org.eclipse.jetty.servlet.ServletHandler.doHandle( > ServletHan > >>>> dler.java:582) > >>>> at org.eclipse.jetty.server.handler.ScopedHandler.handle( > Scoped > >>>> Handler.java:143) > >>>> at org.eclipse.jetty.security.SecurityHandler.handle( > SecurityHa > >>>> ndler.java:548) > >>>> at org.eclipse.jetty.server.session.SessionHandler.doHandle( > >>>> SessionHandler.java:226) > >>>> at org.eclipse.jetty.server.handler.ContextHandler.doHandle( > >>>> ContextHandler.java:1180) > >>>> at org.eclipse.jetty.servlet.ServletHandler.doScope( > ServletHand > >>>> ler.java:512) > >>>> at org.eclipse.jetty.server.session.SessionHandler.doScope( > >>>> SessionHandler.java:185) > >>>> at org.eclipse.jetty.server.handler.ContextHandler.doScope( > >>>> ContextHandler.java:1112) > >>>> at org.eclipse.jetty.server.handler.ScopedHandler.handle( > Scoped > >>>> Handler.java:141) > >>>> at org.eclipse.jetty.server.handler. > ContextHandlerCollection.ha > >>>> ndle(ContextHandlerCollection.java:213) > >>>> at org.eclipse.jetty.server.handler.HandlerCollection. > handle( > >>>> HandlerCollection.java:119) > >>>> at org.eclipse.jetty.server.handler.HandlerWrapper.handle( > Handl > >>>> erWrapper.java:134) > >>>> at org.eclipse.jetty.rewrite.handler.RewriteHandler.handle( > Rewr > >>>> iteHandler.java:335) > >>>> at org.eclipse.jetty.server.handler.HandlerWrapper.handle( > Handl > >>>> erWrapper.java:134) > >>>> at org.eclipse.jetty.server.Server.handle(Server.java:534) > >>>> at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel. > >>>> java:320) > >>>> at org.eclipse.jetty.server.HttpConnection.onFillable( > HttpConne > >>>> ction.java:251) > >>>> at org.eclipse.jetty.io.AbstractConnection$ReadCallback. > >>>> succeeded(AbstractConnection.java:273) > >>>> at org.eclipse.jetty.io.FillInterest.fillable(FillInterest. > >>>> java:95) > >>>> at org.eclipse.jetty.io.SelectChannelEndPoint$2.run( > SelectChann > >>>> elEndPoint.java:93) > >>>> at org.eclipse.jetty.util.thread. > strategy.ExecuteProduceConsume > >>>> .executeProduceConsume(ExecuteProduceConsume.java:303) > >>>> at org.eclipse.jetty.util.thread. > strategy.ExecuteProduceConsume > >>>> .produceConsume(ExecuteProduceConsume.java:148) > >>>> at org.eclipse.jetty.util.thread. > strategy.ExecuteProduceConsume > >>>> .run(ExecuteProduceConsume.java:136) > >>>> at org.eclipse.jetty.util.thread. > QueuedThreadPool.runJob(Queued > >>>> ThreadPool.java:671) > >>>> at org.eclipse.jetty.util.thread. > QueuedThreadPool$2.run(QueuedT > >>>> hreadPool.java:589) > >>>> at java.lang.Thread.run(Thread.java:748) > >>>> Caused by: org.apache.hadoop.ipc.RemoteException(java.io. > IOException): > >>>> File /solr6.6.0/UNCLASS/core_node108/data/tlog/tlog. > 0000000000000006211 > >>>> could only be replicated to 0 nodes instead of minReplication (=1). > >>>> There > >>>> are 40 datanode(s) running and no node(s) are excluded in this > >>>> operation. > >>>> at org.apache.hadoop.hdfs.server. > blockmanagement.BlockManager.c > >>>> hooseTarget4NewBlock(BlockManager.java:1622) > >>>> at org.apache.hadoop.hdfs.server. > namenode.FSNamesystem.getAddit > >>>> ionalBlock(FSNamesystem.java:3351) > >>>> at org.apache.hadoop.hdfs.server. > namenode.NameNodeRpcServer.add > >>>> Block(NameNodeRpcServer.java:683) > >>>> at org.apache.hadoop.hdfs.server. > namenode.AuthorizationProvider > >>>> ProxyClientProtocol.addBlock(AuthorizationProviderProxyClien > >>>> tProtocol.java:214) > >>>> at org.apache.hadoop.hdfs.protocolPB. > ClientNamenodeProtocolServ > >>>> erSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSide > >>>> TranslatorPB.java:495) > >>>> at org.apache.hadoop.hdfs.protocol.proto. > ClientNamenodeProtocol > >>>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam > >>>> enodeProtocolProtos.java) > >>>> at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ > ProtoBufRpcIn > >>>> voker.call(ProtobufRpcEngine.java:617) > >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) > >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java: > 2216) > >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java: > 2212) > >>>> at java.security.AccessController.doPrivileged(Native > Method) > >>>> at javax.security.auth.Subject.doAs(Subject.java:422) > >>>> at org.apache.hadoop.security.UserGroupInformation.doAs( > UserGro > >>>> upInformation.java:1920) > >>>> at org.apache.hadoop.ipc.Server$ > Handler.run(Server.java:2210) > >>>> > >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1475) > >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1412) > >>>> at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke( > >>>> ProtobufRpcEngine.java:229) > >>>> at com.sun.proxy.$Proxy11.addBlock(Unknown Source) > >>>> at org.apache.hadoop.hdfs.protocolPB. > ClientNamenodeProtocolTran > >>>> slatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:418) > >>>> at sun.reflect.GeneratedMethodAccessor18.invoke(Unknown > Source) > >>>> at sun.reflect.DelegatingMethodAccessorImpl. > invoke(DelegatingMe > >>>> thodAccessorImpl.java:43) > >>>> at java.lang.reflect.Method.invoke(Method.java:498) > >>>> at org.apache.hadoop.io.retry.RetryInvocationHandler. > invokeMeth > >>>> od(RetryInvocationHandler.java:191) > >>>> at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke( > Ret > >>>> ryInvocationHandler.java:102) > >>>> at com.sun.proxy.$Proxy12.addBlock(Unknown Source) > >>>> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer. > locateFo > >>>> llowingBlock(DFSOutputStream.java:1459) > >>>> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer. > nextBloc > >>>> kOutputStream(DFSOutputStream.java:1255) > >>>> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run( > >>>> DFSOutputStream.java:449) > >>>> > >>>> 2017-07-17 12:29:24.164 ERROR (qtp985934102-5161057) [c:UNCLASS > >>>> s:shard43 > >>>> r:core_node108 x:UNCLASS_shard43_replica1] o.a.s.s.HttpSolrCall null: > >>>> org.apache.solr.common.SolrException: Error logging add > >>>> at org.apache.solr.update.TransactionLog.write( > TransactionLog. > >>>> java:418) > >>>> at org.apache.solr.update.UpdateLog.add(UpdateLog.java:532) > >>>> at org.apache.solr.update.UpdateLog.add(UpdateLog.java:516) > >>>> at org.apache.solr.update.processor. > DistributedUpdateProcessor. > >>>> versionAdd(DistributedUpdateProcessor.java:1113) > >>>> at org.apache.solr.update.processor. > DistributedUpdateProcessor. > >>>> processAdd(DistributedUpdateProcessor.java:748) > >>>> at org.apache.solr.update.processor. > LogUpdateProcessorFactory$L > >>>> ogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:103) > >>>> at org.apache.solr.handler.loader.JavabinLoader$1.update( > Javabi > >>>> nLoader.java:98) > >>>> at org.apache.solr.client.solrj.request. > JavaBinUpdateRequestCod > >>>> ec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:180) > >>>> at org.apache.solr.client.solrj.request. > JavaBinUpdateRequestCod > >>>> ec$1.readIterator(JavaBinUpdateRequestCodec.java:136) > >>>> at org.apache.solr.common.util.JavaBinCodec.readObject( > JavaBinC > >>>> odec.java:306) > >>>> at org.apache.solr.common.util.JavaBinCodec.readVal( > JavaBinCode > >>>> c.java:251) > >>>> at org.apache.solr.client.solrj.request. > JavaBinUpdateRequestCod > >>>> ec$1.readNamedList(JavaBinUpdateRequestCodec.java:122) > >>>> at org.apache.solr.common.util.JavaBinCodec.readObject( > JavaBinC > >>>> odec.java:271) > >>>> at org.apache.solr.common.util.JavaBinCodec.readVal( > JavaBinCode > >>>> c.java:251) > >>>> at org.apache.solr.common.util.JavaBinCodec.unmarshal( > JavaBinCo > >>>> dec.java:173) > >>>> at org.apache.solr.client.solrj.request. > JavaBinUpdateRequestCod > >>>> ec.unmarshal(JavaBinUpdateRequestCodec.java:187) > >>>> at org.apache.solr.handler.loader.JavabinLoader. > parseAndLoadDoc > >>>> s(JavabinLoader.java:108) > >>>> at org.apache.solr.handler.loader.JavabinLoader.load( > JavabinLoa > >>>> der.java:55) > >>>> at org.apache.solr.handler.UpdateRequestHandler$1.load( > UpdateRe > >>>> questHandler.java:97) > >>>> at org.apache.solr.handler.ContentStreamHandlerBase. > handleReque > >>>> stBody(ContentStreamHandlerBase.java:68) > >>>> at org.apache.solr.handler.RequestHandlerBase. > handleRequest(Req > >>>> uestHandlerBase.java:173) > >>>> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2477) > >>>> at org.apache.solr.servlet.HttpSolrCall.execute( > HttpSolrCall. > >>>> java:723) > >>>> at org.apache.solr.servlet.HttpSolrCall.call( > HttpSolrCall.java: > >>>> 529) > >>>> at org.apache.solr.servlet.SolrDispatchFilter.doFilter( > SolrDisp > >>>> atchFilter.java:361) > >>>> at org.apache.solr.servlet.SolrDispatchFilter.doFilter( > SolrDisp > >>>> atchFilter.java:305) > >>>> at org.eclipse.jetty.servlet.ServletHandler$CachedChain. > doFilte > >>>> r(ServletHandler.java:1691) > >>>> at org.eclipse.jetty.servlet.ServletHandler.doHandle( > ServletHan > >>>> dler.java:582) > >>>> at org.eclipse.jetty.server.handler.ScopedHandler.handle( > Scoped > >>>> Handler.java:143) > >>>> at org.eclipse.jetty.security.SecurityHandler.handle( > SecurityHa > >>>> ndler.java:548) > >>>> at org.eclipse.jetty.server.session.SessionHandler.doHandle( > >>>> SessionHandler.java:226) > >>>> at org.eclipse.jetty.server.handler.ContextHandler.doHandle( > >>>> ContextHandler.java:1180) > >>>> at org.eclipse.jetty.servlet.ServletHandler.doScope( > ServletHand > >>>> ler.java:512) > >>>> at org.eclipse.jetty.server.session.SessionHandler.doScope( > >>>> SessionHandler.java:185) > >>>> at org.eclipse.jetty.server.handler.ContextHandler.doScope( > >>>> ContextHandler.java:1112) > >>>> at org.eclipse.jetty.server.handler.ScopedHandler.handle( > Scoped > >>>> Handler.java:141) > >>>> at org.eclipse.jetty.server.handler. > ContextHandlerCollection.ha > >>>> ndle(ContextHandlerCollection.java:213) > >>>> at org.eclipse.jetty.server.handler.HandlerCollection. > handle( > >>>> HandlerCollection.java:119) > >>>> at org.eclipse.jetty.server.handler.HandlerWrapper.handle( > Handl > >>>> erWrapper.java:134) > >>>> at org.eclipse.jetty.rewrite.handler.RewriteHandler.handle( > Rewr > >>>> iteHandler.java:335) > >>>> at org.eclipse.jetty.server.handler.HandlerWrapper.handle( > Handl > >>>> erWrapper.java:134) > >>>> at org.eclipse.jetty.server.Server.handle(Server.java:534) > >>>> at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel. > >>>> java:320) > >>>> at org.eclipse.jetty.server.HttpConnection.onFillable( > HttpConne > >>>> ction.java:251) > >>>> at org.eclipse.jetty.io.AbstractConnection$ReadCallback. > >>>> succeeded(AbstractConnection.java:273) > >>>> at org.eclipse.jetty.io.FillInterest.fillable(FillInterest. > >>>> java:95) > >>>> at org.eclipse.jetty.io.SelectChannelEndPoint$2.run( > SelectChann > >>>> elEndPoint.java:93) > >>>> at org.eclipse.jetty.util.thread. > strategy.ExecuteProduceConsume > >>>> .executeProduceConsume(ExecuteProduceConsume.java:303) > >>>> at org.eclipse.jetty.util.thread. > strategy.ExecuteProduceConsume > >>>> .produceConsume(ExecuteProduceConsume.java:148) > >>>> at org.eclipse.jetty.util.thread. > strategy.ExecuteProduceConsume > >>>> .run(ExecuteProduceConsume.java:136) > >>>> at org.eclipse.jetty.util.thread. > QueuedThreadPool.runJob(Queued > >>>> ThreadPool.java:671) > >>>> at org.eclipse.jetty.util.thread. > QueuedThreadPool$2.run(QueuedT > >>>> hreadPool.java:589) > >>>> at java.lang.Thread.run(Thread.java:748) > >>>> Caused by: org.apache.hadoop.ipc.RemoteException(java.io. > IOException): > >>>> File /solr6.6.0/UNCLASS/core_node108/data/tlog/tlog. > 0000000000000006211 > >>>> could only be replicated to 0 nodes instead of minReplication (=1). > >>>> There > >>>> are 40 datanode(s) running and no node(s) are excluded in this > >>>> operation. > >>>> at org.apache.hadoop.hdfs.server. > blockmanagement.BlockManager.c > >>>> hooseTarget4NewBlock(BlockManager.java:1622) > >>>> at org.apache.hadoop.hdfs.server. > namenode.FSNamesystem.getAddit > >>>> ionalBlock(FSNamesystem.java:3351) > >>>> at org.apache.hadoop.hdfs.server. > namenode.NameNodeRpcServer.add > >>>> Block(NameNodeRpcServer.java:683) > >>>> at org.apache.hadoop.hdfs.server. > namenode.AuthorizationProvider > >>>> ProxyClientProtocol.addBlock(AuthorizationProviderProxyClien > >>>> tProtocol.java:214) > >>>> at org.apache.hadoop.hdfs.protocolPB. > ClientNamenodeProtocolServ > >>>> erSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSide > >>>> TranslatorPB.java:495) > >>>> at org.apache.hadoop.hdfs.protocol.proto. > ClientNamenodeProtocol > >>>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam > >>>> enodeProtocolProtos.java) > >>>> at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ > ProtoBufRpcIn > >>>> voker.call(ProtobufRpcEngine.java:617) > >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) > >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java: > 2216) > >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java: > 2212) > >>>> at java.security.AccessController.doPrivileged(Native > Method) > >>>> at javax.security.auth.Subject.doAs(Subject.java:422) > >>>> at org.apache.hadoop.security.UserGroupInformation.doAs( > UserGro > >>>> upInformation.java:1920) > >>>> at org.apache.hadoop.ipc.Server$ > Handler.run(Server.java:2210) > >>>> > >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1475) > >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1412) > >>>> at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke( > >>>> ProtobufRpcEngine.java:229) > >>>> at com.sun.proxy.$Proxy11.addBlock(Unknown Source) > >>>> at org.apache.hadoop.hdfs.protocolPB. > ClientNamenodeProtocolTran > >>>> slatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:418) > >>>> at sun.reflect.GeneratedMethodAccessor18.invoke(Unknown > Source) > >>>> at sun.reflect.DelegatingMethodAccessorImpl. > invoke(DelegatingMe > >>>> thodAccessorImpl.java:43) > >>>> at java.lang.reflect.Method.invoke(Method.java:498) > >>>> at org.apache.hadoop.io.retry.RetryInvocationHandler. > invokeMeth > >>>> od(RetryInvocationHandler.java:191) > >>>> at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke( > Ret > >>>> ryInvocationHandler.java:102) > >>>> at com.sun.proxy.$Proxy12.addBlock(Unknown Source) > >>>> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer. > locateFo > >>>> llowingBlock(DFSOutputStream.java:1459) > >>>> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer. > nextBloc > >>>> kOutputStream(DFSOutputStream.java:1255) > >>>> at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run( > >>>> DFSOutputStream.java:449) > >>>> > >>>> 2017-07-17 12:29:24.187 INFO > >>>> (zkCallback-5-thread-144-processing-n:juliet:9100_solr) > >>>> [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent > >>>> state:SyncConnected type:NodeDataChanged > >>>> path:/collections/UNCLASS/state.json] > >>>> for collection [UNCLASS] has occurred - updating... (live nodes size: > >>>> [45]) > >>>> > >>>> On the client side, the error looks like: > >>>> 2017-07-16 19:03:16,118 WARN > >>>> [com.ngc.bigdata.ie_solrindexer.IndexDocument] > >>>> Indexing error: org.apache.solr.client.solrj.i > >>>> mpl.CloudSolrClient$RouteException: Error from server at > >>>> http://leda:9100/solr/UNCLASS_shard44_replica3: Exception writing > >>>> document id COLLECT10086453202 to the index; possible analysis error. > >>>> for > >>>> collection: UNCLASS > >>>> org.apache.solr.client.solrj.impl.CloudSolrClient$RouteException: > Error > >>>> from server at http://leda:9100/solr/UNCLASS_shard44_replica3: > Exception > >>>> writing document id COLLECT10086453202 to the index; possible analysis > >>>> error. > >>>> at org.apache.solr.client.solrj.impl.CloudSolrClient. > directUpda > >>>> te(CloudSolrClient.java:819) > >>>> at org.apache.solr.client.solrj.impl.CloudSolrClient. > sendReques > >>>> t(CloudSolrClient.java:1263) > >>>> at org.apache.solr.client.solrj.impl.CloudSolrClient. > requestWit > >>>> hRetryOnStaleState(CloudSolrClient.java:1134) > >>>> at org.apache.solr.client.solrj. > impl.CloudSolrClient.request(Cl > >>>> oudSolrClient.java:1073) > >>>> at org.apache.solr.client.solrj.SolrRequest.process( > SolrRequest > >>>> .java:160) > >>>> at org.apache.solr.client.solrj.SolrClient.add(SolrClient. > java: > >>>> 106) > >>>> at org.apache.solr.client.solrj.SolrClient.add(SolrClient. > java: > >>>> 71) > >>>> at org.apache.solr.client.solrj.SolrClient.add(SolrClient. > java: > >>>> 85) > >>>> at com.ngc.bigdata.ie_solrindexer.IndexDocument. > indexSolrDocs( > >>>> IndexDocument.java:959) > >>>> at com.ngc.bigdata.ie_solrindexer.IndexDocument.index( > >>>> IndexDocument.java:236) > >>>> at com.ngc.bigdata.ie_solrindexer. > SolrIndexerProcessor.doWork(S > >>>> olrIndexerProcessor.java:63) > >>>> at com.ngc.intelenterprise.intelentutil.utils.Processor.run( > >>>> Processor.java:140) > >>>> at com.ngc.intelenterprise.intelentutil.jms. > IntelEntQueueProc. > >>>> process(IntelEntQueueProc.java:208) > >>>> at org.apache.camel.processor.DelegateSyncProcessor.process( > Del > >>>> egateSyncProcessor.java:63) > >>>> at org.apache.camel.management.InstrumentationProcessor. > process > >>>> (InstrumentationProcessor.java:77) > >>>> at org.apache.camel.processor.RedeliveryErrorHandler. > process(Re > >>>> deliveryErrorHandler.java:460) > >>>> at org.apache.camel.processor.CamelInternalProcessor. > process(Ca > >>>> melInternalProcessor.java:190) > >>>> at org.apache.camel.processor.CamelInternalProcessor. > process(Ca > >>>> melInternalProcessor.java:190) > >>>> at org.apache.camel.component.seda.SedaConsumer. > sendToConsumers > >>>> (SedaConsumer.java:298) > >>>> at org.apache.camel.component.seda.SedaConsumer.doRun( > SedaConsu > >>>> mer.java:207) > >>>> at org.apache.camel.component.seda.SedaConsumer.run( > SedaConsume > >>>> r.java:154) > >>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPool > >>>> Executor.java:1142) > >>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoo > >>>> lExecutor.java:617) > >>>> at java.lang.Thread.run(Thread.java:748) > >>>> Caused by: > >>>> org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: > >>>> Error from server at http://leda:9100/solr/UNCLASS_shard44_replica3: > >>>> Exception writing document id COLLECT10086453202 to the index; > possible > >>>> analysis error. > >>>> at org.apache.solr.client.solrj.impl.HttpSolrClient. > executeMeth > >>>> od(HttpSolrClient.java:610) > >>>> at org.apache.solr.client.solrj.impl.HttpSolrClient.request( > Htt > >>>> pSolrClient.java:279) > >>>> at org.apache.solr.client.solrj.impl.HttpSolrClient.request( > Htt > >>>> pSolrClient.java:268) > >>>> at org.apache.solr.client.solrj.impl.LBHttpSolrClient. > doRequest > >>>> (LBHttpSolrClient.java:447) > >>>> at org.apache.solr.client.solrj. > impl.LBHttpSolrClient.request(L > >>>> BHttpSolrClient.java:388) > >>>> at org.apache.solr.client.solrj.impl.CloudSolrClient.lambda$ > dir > >>>> ectUpdate$0(CloudSolrClient.java:796) > >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) > >>>> at org.apache.solr.common.util.ExecutorUtil$ > MDCAwareThreadPoolE > >>>> xecutor.lambda$execute$0(ExecutorUtil.java:229) > >>>> ... 3 more > >>>> 2017-07-16 19:03:16,134 ERROR > >>>> [com.ngc.bigdata.ie_solrindexer.IndexDocument] > >>>> Error indexing: org.apache.solr.client.solrj.i > >>>> mpl.CloudSolrClient$RouteException: Error from server at > >>>> http://leda:9100/solr/UNCLASS_shard44_replica3: Exception writing > >>>> document id COLLECT10086453202 to the index; possible analysis error. > >>>> for > >>>> collection: UNCLASS. > >>>> 2017-07-16 19:03:16,135 ERROR > >>>> [com.ngc.bigdata.ie_solrindexer.IndexDocument] > >>>> Exception during indexing: org.apache.solr.client.solrj.i > >>>> mpl.CloudSolrClient$RouteException: Error from server at > >>>> http://leda:9100/solr/UNCLASS_shard44_replica3: Exception writing > >>>> document id COLLECT10086453202 to the index; possible analysis error. > >>>> > >>>> I can fire them back up, but they only run for a short time before > >>>> getting more indexing errors. Several of the nodes show as down in > the > >>>> cloud view. Any help would be appreciated! Thank you! > >>>> > >>>> > >>>> -Joe > >>>> > >>>> > >> > >> --- > >> This email has been checked for viruses by AVG. > >> http://www.avg.com > >> > > >