Garbage Collectors Summary: https://apps.sematext.com/spm-reports/s/rgRnwuShgI
Pool Size: https://apps.sematext.com/spm-reports/s/H16ndqichM First Stopping recovery warning: 4:00, OOM error: 6:30. 2014-03-24 16:35 GMT+02:00 Shalin Shekhar Mangar <shalinman...@gmail.com>: > I am guessing that it is all related to memory issues. I guess that as > the used heap increases, full GC cycles increase causing ZK timeouts > which in turn cause more recoveries to be initiated. In the end, > everything blows up with the out of memory errors. Do you log GC > activity on your servers? > > I suggest that you rollback to 4.6.1 for now and upgrade to 4.7.1 when > it releases next week. > > On Mon, Mar 24, 2014 at 7:51 PM, Lukas Mikuckis <lukasmikuc...@gmail.com> > wrote: > > Yes, we upgraded solr from 4.6.1 to 4.7 3 weeks ago (2 weeks before solr > > started crashing). > > When we were upgrading, we just upgraded solr and changed versions in > > collections configs. > > > > When solr crashes we get OOM but only 2h after first Stopping recovery > > warnings. > > > > Maybe you have any ideas when Stopping recovery warnings are thrown? > > Because now we have no idea what could cause this issue. > > > > Mon, 24 Mar 2014 04:03:17 GMT Shalin Shekhar Mangar < > shalinman...@gmail.com > >>: > >> > >> Did you upgrade recently to Solr 4.7? 4.7 has a bad bug which can > >> cause out of memory issues. Can you check your logs for out of memory > >> errors? > >> > >> On Sun, Mar 23, 2014 at 9:07 PM, Lukas Mikuckis < > lukasmikuc...@gmail.com> > > wrote: > >> > Solr version: 4.7 > >> > > >> > Architecture: > >> > 2 solrs (1 shard, leader + replica) > >> > 3 zookeepers > >> > > >> > Servers: > >> > * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores > >> > * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores > >> > * zookeeper > >> > > >> > Solr data: > >> > * 21 collections > >> > * Many fields, small docs, docs count per collection from 1k to 500k > >> > > >> > About a week ago solr started crashing. It crashes every day, 3-4 > times > > a > >> > day. Usually at nigh. I can't tell anything what could it be related > to > >> > because at that time we haven't done any configuration changes. Load > >> > haven't changed too. > >> > > >> > > >> > Everything starts with Stopping recovery for .. warnings (every > > warnings is > >> > repeated several times): > >> > > >> > WARN org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for > >> > zkNodeName=core_node1core=****************** > >> > > >> > WARN org.apache.solr.cloud.ElectionContext; cancelElection did not > find > >> > election node to remove > >> > > >> > WARN org.apache.solr.update.PeerSync; no frame of reference to tell > if > >> > we've missed updates > >> > > >> > WARN - 2014-03-23 04:00:26.286; org.apache.solr.update.PeerSync; no > > frame > >> > of reference to tell if we've missed updates > >> > > >> > WARN - 2014-03-23 04:00:30.728; org.apache.solr.handler.SnapPuller; > > File > >> > _f9m_Lucene41_0.doc expected to be 6218278 while it is 7759879 > >> > > >> > WARN - 2014-03-23 04:00:54.126; > >> > org.apache.solr.update.UpdateLog$LogReplayer; Starting log replay > >> > > > > tlog{file=/path/solr/collection1_shard1_replica2/data/tlog/tlog.0000000000000003272 > >> > refcount=2} active=true starting pos=356216606 > >> > > >> > Then again Stopping recovery for .. warnings: > >> > > >> > WARN org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for > >> > zkNodeName=core_node1core=****************** > >> > > >> > ERROR - 2014-03-23 05:19:29.566; org.apache.solr.common.SolrException; > >> > org.apache.solr.common.SolrException: No registered leader was found > > after > >> > waiting for 4000ms , collection: collection1 slice: shard1 > >> > > >> > ERROR - 2014-03-23 05:20:03.961; org.apache.solr.common.SolrException; > >> > org.apache.solr.common.SolrException: I was asked to wait on state > down > > for > >> > IP:PORT_solr but I still do not see the requested state. I see state: > >> > active live:false > >> > > >> > > >> > After this serves mostly didn't recover. > >> > >> > >> > >> -- > >> Regards, > >> Shalin Shekhar Mangar. > >> > >> > > > > -- > Regards, > Shalin Shekhar Mangar. >