Delay in replication between cloud servers

2016-02-16 Thread Cool Techi
We are using solr cloud with 1 shard and replication factor as 3. We are 
noticing that the time for data to become available across all replicas from 
the leader is very high.
The data rate is not very high, is there anyway to control this. In 
master-slave setup with give a replication time.
Regards,Rohit  
  

RE: Delay in replication between cloud servers

2016-02-16 Thread Cool Techi
Further we have noticed that the delay increase a couple of hours after 
restart. Details related to sorlconfig.xml are given below,
  
   15000 
   25000
   false 
 

 
 1000 
   Regards,Rohit 

> From: cooltec...@outlook.com
> To: solr-user@lucene.apache.org
> Subject: Delay in replication between cloud servers
> Date: Tue, 16 Feb 2016 20:20:04 +0530
> 
> We are using solr cloud with 1 shard and replication factor as 3. We are 
> noticing that the time for data to become available across all replicas from 
> the leader is very high.
> The data rate is not very high, is there anyway to control this. In 
> master-slave setup with give a replication time.
> Regards,Rohit  
> 
  

RE: Delay in replication between cloud servers

2016-02-21 Thread Cool Techi
Hi,
Can someone point in the right direction to check this.
Regards,Rohit 

> From: cooltec...@outlook.com
> To: solr-user@lucene.apache.org
> Subject: RE: Delay in replication between cloud servers
> Date: Tue, 16 Feb 2016 22:16:23 +0530
> 
> Further we have noticed that the delay increase a couple of hours after 
> restart. Details related to sorlconfig.xml are given below,
>   
>15000 
>25000
>false 
>  
> 
>  
>  1000 
>Regards,Rohit 
> 
> > From: cooltec...@outlook.com
> > To: solr-user@lucene.apache.org
> > Subject: Delay in replication between cloud servers
> > Date: Tue, 16 Feb 2016 20:20:04 +0530
> > 
> > We are using solr cloud with 1 shard and replication factor as 3. We are 
> > noticing that the time for data to become available across all replicas 
> > from the leader is very high.
> > The data rate is not very high, is there anyway to control this. In 
> > master-slave setup with give a replication time.
> > Regards,Rohit  
> >   
> 
  

Solr Cloud leader propagates slow to other nodes

2016-03-04 Thread Cool Techi
Hi,
We are using solr  4.8.1 cloud and have a single shard core and 3 replicas in 
total.  We are noticing the following issues,
1) Master propagates slow to to the other nodes/replicas.2) This increase over 
a period of time, i.e the data propagation time increase as the time from which 
leader is elected increases? 
What could be causing this, our commit settings are given below.
  
   12000 
   25000
   false 
 



 1000 
   
Regards,Ayush 









  

RE: Solr Cloud leader propagates slow to other nodes

2016-03-04 Thread Cool Techi
Any suggestion here would be highly appreciated, this is hampering our 
production performance and we are not able to pin point any reason.




> From: cooltec...@outlook.com
> To: solr-user@lucene.apache.org
> Subject: Solr Cloud leader propagates slow to other nodes
> Date: Fri, 4 Mar 2016 18:16:07 +0530
> 
> Hi,
> We are using solr  4.8.1 cloud and have a single shard core and 3 replicas in 
> total.  We are noticing the following issues,
> 1) Master propagates slow to to the other nodes/replicas.2) This increase 
> over a period of time, i.e the data propagation time increase as the time 
> from which leader is elected increases? 
> What could be causing this, our commit settings are given below.
>   
>12000 
>25000
>false 
>  
> 
> 
> 
>  1000 
>
> Regards,Ayush 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
  

Search in URL

2014-03-28 Thread Cool Techi
Hi,
We have several tweets stored in our system which are of the following nature,
https://twitter.com/SharFairies/status/448992742257881088
I am searching for keywords like starcinema or abs-cbn.com, but these results 
are not returned, how can I search in this setup. My field type is defined as 
below,
  

 
   

 
 


  

Fails to index if unique field has special characters

2014-04-10 Thread Cool Techi
Hi,
We are migrating from Solr 4.6 standalone to Solr 4.7 cloud version, while 
reindexing the document we are getting the following error. This is happening 
when the unique key has special character, this was not noticed in version 4.6 
standalone mode, so we are not sure if this is a version problem or a cloud 
issue. Example of the unique key is given below,
http://www.mynews.in/Blog/smrity!!**)))!miami_dolphins_vs_dallas_cowboys_live_stream_on_line_nfl_football_free_video_broadcast_B142707.html
Exception Stack Trace
ERROR - 2014-04-10 10:51:44.361; org.apache.solr.common.SolrException; 
java.lang.ArrayIndexOutOfBoundsException: 2   at 
org.apache.solr.common.cloud.CompositeIdRouter$KeyParser.getHash(CompositeIdRouter.java:296)
   at 
org.apache.solr.common.cloud.CompositeIdRouter.sliceHash(CompositeIdRouter.java:58)
   at 
org.apache.solr.common.cloud.HashBasedRouter.getTargetSlice(HashBasedRouter.java:33)
   at 
org.apache.solr.update.processor.DistributedUpdateProcessor.setupRequest(DistributedUpdateProcessor.java:218)
   at 
org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:550)
   at 
org.apache.solr.update.processor.LogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:100)
   at 
org.apache.solr.handler.loader.XMLLoader.processUpdate(XMLLoader.java:247)  
 at org.apache.solr.handler.loader.XMLLoader.load(XMLLoader.java:174)   at 
org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:92)
   at 
org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:74)
   at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
   at org.apache.solr.core.SolrCore.execute(SolrCore.java:1916)   at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:780) 
  at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:427)
   at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:217)
   at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
   at 
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)  
 at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)   
at 
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557) 
  at org.eclipse.jetty.server.session.SessionHandle

Thanks,Ayush  

RE: Fails to index if unique field has special characters

2014-04-11 Thread Cool Techi

Thanks, that was helpful.
Regards,Rohit
> Date: Thu, 10 Apr 2014 08:44:36 -0700
> From: iori...@yahoo.com
> Subject: Re: Fails to index if unique field has special characters
> To: solr-user@lucene.apache.org
> 
> Hi Ayush,
> 
> I thinks this 
> 
> ""IBM!12345". The exclamation mark ('!') is critical here, as it 
> distinguishes the prefix used to determine which shard to direct the document 
> to."
> 
> https://cwiki.apache.org/confluence/display/solr/Shards+and+Indexing+Data+in+SolrCloud
> 
> 
> 
> 
> On Thursday, April 10, 2014 2:35 PM, Cool Techi  
> wrote:
> Hi,
> We are migrating from Solr 4.6 standalone to Solr 4.7 cloud version, while 
> reindexing the document we are getting the following error. This is happening 
> when the unique key has special character, this was not noticed in version 
> 4.6 standalone mode, so we are not sure if this is a version problem or a 
> cloud issue. Example of the unique key is given below,
> http://www.mynews.in/Blog/smrity!!**)))!miami_dolphins_vs_dallas_cowboys_live_stream_on_line_nfl_football_free_video_broadcast_B142707.html
> Exception Stack Trace
> ERROR - 2014-04-10 10:51:44.361; org.apache.solr.common.SolrException; 
> java.lang.ArrayIndexOutOfBoundsException: 2   at 
> org.apache.solr.common.cloud.CompositeIdRouter$KeyParser.getHash(CompositeIdRouter.java:296)
>at 
> org.apache.solr.common.cloud.CompositeIdRouter.sliceHash(CompositeIdRouter.java:58)
>at 
> org.apache.solr.common.cloud.HashBasedRouter.getTargetSlice(HashBasedRouter.java:33)
>at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.setupRequest(DistributedUpdateProcessor.java:218)
>at 
> org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:550)
>at 
> org.apache.solr.update.processor.LogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:100)
>at 
> org.apache.solr.handler.loader.XMLLoader.processUpdate(XMLLoader.java:247)
>at org.apache.solr.handler.loader.XMLLoader.load(XMLLoader.java:174)   
> at
>  
> org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:92)
>at 
> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:74)
>at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
>at org.apache.solr.core.SolrCore.execute(SolrCore.java:1916)   at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:780)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:427)
>at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:217)
>at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
>at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
>at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137) 
>   at
>  org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)  
>  at org.eclipse.jetty.server.session.SessionHandle
> 
> Thanks,Ayush   
  

RE: Inconsistent response from Cloud Query

2014-05-14 Thread Cool Techi
We have noticed Solr returns in-consistent results during replica recovery and 
not all replicas are in the same state, so when your query goes to a replica 
which might be recovering or still copying the index then the counts may differ.
regards,Ayush

> Date: Tue, 6 May 2014 16:14:36 +0530
> Subject: Inconsistent response from Cloud Query
> From: clearmido...@gmail.com
> To: solr-user@lucene.apache.org
> 
> Hi All,
> 
> I have setup cloud-4.6.2 with default configuration on single machine with
> 2 shards and 2 replication through
> https://cwiki.apache.org/confluence/display/solr/Getting+Started+with+SolrCloud
> 
> Cloud was up and running and I indexed the example data xml to it, it went
> fine.
> Now when I am querying with *distrib=true* it is giving inconsistent
> result, sometimes it gives 4 Result Response and sometimes 8(Actual Number)
> 
> Has anyone been through the situation. Looking for positive and quick
> response.
> 
> Thanks!
  

Storing tweets For WC2014

2014-05-15 Thread Cool Techi
Hi,
We have a requirement from one of our customers to provide search and analytics 
on the upcoming Soccer World cup, given the sheer volume of tweet's that would 
be generated at such an event I cannot imagine what would be required to store 
this in solr. 
It would be great if there can be some pointer's on the scale or hardware 
required, number of shards that should be created etc. Some requirement,
All the tweets should be searchable (approximately 100million tweets/date  * 60 
Days of event). All fields on tweets should be searchable/facet on numeric and 
date fields. Facets would be run on TwitterId's (unique users), tweet created 
on date, Location, Sentiment (some fields which we generate)

If anyone has attempted anything like this it would be helpful.
Regards,Rohit
  

Inconsistant result's on solr cloud 4.8

2014-07-16 Thread Cool Techi
Hi,
We are using solr cloud with solr version 4.8, we have 2 shard/2 replica 
servers in Solr Cloud. During two consecutive request to the solr cloud, the 
"total results" number varies. 
1) As per my understanding this can happen when the leader and the replica have 
inconsistant number of results.
2) This inconsistant number of docs between leader and replica can happen only 
when replica is recovering. Should a request be sent to a node which is 
recovering.
Since this is happening on our live setup, we tend to question how much can we 
rely on solr. What could be causing this and what's the fix.
Regards   

SolrCloud Issues

2014-07-17 Thread Cool Techi
Hi,
We have moved to SolrCloud (version 4.8) and are facing several issues in this 
setup compared to the master/slave setup we have had for a while now,
1) We have a 2 shard set-up with one replica each, we notice that most of the 
time the replica's are in recovering status. What could be causing this?
2) In-consistant results count, especially when one of the nodes is recovering, 
I have asked another question earlier on this. To our understanding the 
recovering node, doesn't returns any results, what else can cause this?
3) Solr node goes down very frequently, there is no OOM or otherwise in the 
logs, but a node keeps going down. Also, at times we have noticed that tomcat 
stops responding. Since there are so many parts to solr cloud, checking what's 
causing the issue is difficult in a quick time, so if anyone else has faced 
this, it would be very helpful.

Since this is happening on our UAT environment, we need to get a fix soon. 
Regards,Ayush 

RE: SolrCloud Issues

2014-07-17 Thread Cool Techi
1) The Zookeepers are on the same node as Solr, should we move them out? What 
would be the basic config's of machine just running Zookeepers?
2) The servers are pretty big, 
 2 X Quad Core processors 64 bit 96Gb of RAM 500GB SSD 
drive on which solr resides 

The index side between the two shards is 250GB,  how can we detect performance 
problems if any. Also, what keep the node in recovering mode for most of the 
time.  We index about 2K documents/ minute.
What would be the ideal configuration for such load and increasing index size.
Regards,Ayush



> From: dan.mur...@buy4now.com
> To: solr-user@lucene.apache.org
> Subject: RE: SolrCloud Issues
> Date: Thu, 17 Jul 2014 13:27:17 +
> 
> Have you deployed ZooKeeper on servers other than the Solr nodes?
> If you have them on the Solr nodes, then you may be getting elections when 
> under pressure.
> 
> -Original Message-
> From: Shawn Heisey [mailto:s...@elyograg.org] 
> Sent: 17 July 2014 14:25
> To: solr-user@lucene.apache.org
> Subject: Re: SolrCloud Issues
> 
> On 7/17/2014 6:57 AM, Cool Techi wrote:
> > We have moved to SolrCloud (version 4.8) and are facing several issues 
> > in this setup compared to the master/slave setup we have had for a 
> > while now,
> > 1) We have a 2 shard set-up with one replica each, we notice that most of 
> > the time the replica's are in recovering status. What could be causing this?
> > 2) In-consistant results count, especially when one of the nodes is 
> > recovering, I have asked another question earlier on this. To our 
> > understanding the recovering node, doesn't returns any results, what else 
> > can cause this?
> > 3) Solr node goes down very frequently, there is no OOM or otherwise in the 
> > logs, but a node keeps going down. Also, at times we have noticed that 
> > tomcat stops responding. Since there are so many parts to solr cloud, 
> > checking what's causing the issue is difficult in a quick time, so if 
> > anyone else has faced this, it would be very helpful.
> 
> I don't have anything specific for you, but if you are having any kind of 
> performance issues at all, it can lead to bizarre SolrCloud behavior.
> 
> The basic zookeeper client timeout defaults to 15 seconds.  This is a very 
> long timeout, but if anything is happening that makes any part of SolrCloud 
> wait longer than 15 seconds, SolrCloud will think there's a problem that 
> needs recovery.
> 
> Here's a summary of common performance problems and some possible solutions:
> 
> http://wiki.apache.org/solr/SolrPerformanceProblems
> 
> Thanks,
> Shawn
> 
  

Solr Result Tagging

2013-10-25 Thread Cool Techi
Hi,
My search queries to solr are of the following nature,
 (A OR B OR C) OR (X AND Y AND Z) OR ((ABC AND DEF) - XYZ)
What I am trying to achieve is when I fire the query the results returned 
should be able to tagged with which part or the OR resulted in the result. In 
case all three parts above are applicable then the result should indicate the 
same. I tried group.query feature, but doesn't seem like it works on solr cloud.
Thanks,Ayush
  

Faceting within groups

2013-12-07 Thread Cool Techi
Hi,
I am not sure if faceting with groups is supported, the documents do seem to 
suggest it works, but cant seem to get the intended results.
("Amazon Cloud" OR ("IBM Cloud")sourceIdsentimenttruetrue
Also, if it work's does solr cloud support it.
Regards,Ayush 

RE: Faceting within groups

2013-12-08 Thread Cool Techi
Any help here?

> From: cooltec...@outlook.com
> To: solr-user@lucene.apache.org
> Subject: Faceting within groups
> Date: Sat, 7 Dec 2013 14:00:20 +0530
> 
> Hi,
> I am not sure if faceting with groups is supported, the documents do seem to 
> suggest it works, but cant seem to get the intended results.
> ("Amazon Cloud" OR ("IBM Cloud") name="group.field">sourceIdsentiment name="group">truetrue
> Also, if it work's does solr cloud support it.
> Regards,Ayush   
  

Unknow Query exception

2013-12-14 Thread Cool Techi
I am seeing the following exception in solr logs after regular intervals,  we 
are not firing any query to have such error, what could be causing this?










ERROR - 2013-12-14 05:35:52.722; org.apache.solr.common.SolrException; 
org.apache.solr.common.SolrException: org.apache.solr.search.SyntaxError: 
Cannot parse 'docKey:()': Encountered " ")" ") "" at line 1, column 8.
Was expecting one of:
 ...
"+" ...
"-" ...
 ...
"(" ...
"*" ...
 ...
 ...
 ...
 ...
 ...
"[" ...
"{" ...
 ...
 ...
 ...
"*" ...


at 
org.apache.solr.handler.component.QueryComponent.prepare(QueryComponent.java:147)
  

Solr replicating at 5 mb/sec

2014-02-22 Thread Cool Techi
Hi,
I am running solr replication between two machine which are connected by a 1 gb 
network speed. The best speed I am getting for replication is 5mb/sec, how can 
this be increased.
The replication keeps failing and this is the first time replication of an 
index over 300Gb in size. We are using solr verson 4.1 on master and solr 4.3 
on slave.
Regards,Ayush 

Solr 4.3 Master/Slave Issues

2013-07-05 Thread Cool Techi
We have set up solr 4.3 with master/setup and are facing a couple of issues, 

Index locking, the index on slave hangs at time and when we restart the core 
the core get's locked up. I have checked the logs and there are no OOM error or 
anything else other than the error given below,Caused by: 
org.apache.solr.common.SolrException: Index locked for write for core XXX
at org.apache.solr.core.SolrCore.(SolrCore.java:821)
at org.apache.solr.core.SolrCore.(SolrCore.java:618)
at 
org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:949)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:984)
at org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:597)
at org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:592)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
... 1 more
Caused by: org.apache.lucene.store.LockObtainFailedException: Index locked for 
write for core XXat 
org.apache.solr.core.SolrCore.initIndex(SolrCore.java:484)
at org.apache.solr.core.SolrCore.(SolrCore.java:745)
... 13 moreWhen we fire backup on Slave, the whole core hangs after a while 
and also replication stops. This was not happening with our earlier setup of 
solr3.6



Regards,
Ayush













  

RE: Solr 4.3 Master/Slave Issues

2013-07-05 Thread Cool Techi

1) That was my initial suspicion, but when I run ps -aux | grep "java", but 
there it doesn't show any other program running. I kill the process and start 
again and it locks.

2) When we fire backup on Slave, the whole core hangs after a while and also 
replication stops. This was not happening with our earlier setup of solr3.6

Regards,


> Date: Fri, 5 Jul 2013 16:49:24 +0530
> Subject: Re: Solr 4.3 Master/Slave Issues
> From: shalinman...@gmail.com
> To: solr-user@lucene.apache.org
> 
> This can mean multiple things:
> 1. You had killed a solr process earlier which left the lock file in place
> 2. You have more than one Solr core pointing to the same data directory
> 3. A solr process is already running and you are trying to start
> another one with the same config.
> 
> On Fri, Jul 5, 2013 at 2:34 PM, Cool Techi  wrote:
> > We have set up solr 4.3 with master/setup and are facing a couple of issues,
> >
> > Index locking, the index on slave hangs at time and when we restart the 
> > core the core get's locked up. I have checked the logs and there are no OOM 
> > error or anything else other than the error given below,Caused by: 
> > org.apache.solr.common.SolrException: Index locked for write for core XXX
> > at org.apache.solr.core.SolrCore.(SolrCore.java:821)
> > at org.apache.solr.core.SolrCore.(SolrCore.java:618)
> > at 
> > org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:949)
> > at org.apache.solr.core.CoreContainer.create(CoreContainer.java:984)
> > at org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:597)
> > at org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:592)
> > at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
> > at java.util.concurrent.FutureTask.run(FutureTask.java:138)
> > at 
> > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
> > at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
> > at java.util.concurrent.FutureTask.run(FutureTask.java:138)
> > at 
> > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
> > at 
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
> > ... 1 more
> > Caused by: org.apache.lucene.store.LockObtainFailedException: Index locked 
> > for write for core XXat 
> > org.apache.solr.core.SolrCore.initIndex(SolrCore.java:484)
> > at org.apache.solr.core.SolrCore.(SolrCore.java:745)
> > ... 13 moreWhen we fire backup on Slave, the whole core hangs after a 
> > while and also replication stops. This was not happening with our earlier 
> > setup of solr3.6
> >
> >
> >
> > Regards,
> > Ayush
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> 
> 
> 
> -- 
> Regards,
> Shalin Shekhar Mangar.
  

RE: Solr 4.3 Master/Slave Issues

2013-07-05 Thread Cool Techi
The normal tomcat shutdown doesn't stop the server and take a long time, so i 
do issue a kill -9 command. Any other suggestion to do this without the locking.

I would initiate a backup again and send the logs.

regards,
Ayush

> Date: Fri, 5 Jul 2013 19:40:12 +0530
> Subject: Re: Solr 4.3 Master/Slave Issues
> From: shalinman...@gmail.com
> To: solr-user@lucene.apache.org
> 
> On Fri, Jul 5, 2013 at 6:14 PM, Cool Techi  wrote:
> >
> > 1) That was my initial suspicion, but when I run ps -aux | grep "java", but 
> > there it doesn't show any other program running. I kill the process and 
> > start again and it locks.
> 
> How are you killing the process? A SIGKILL will leave a lock file.
> 
> >
> > 2) When we fire backup on Slave, the whole core hangs after a while and 
> > also replication stops. This was not happening with our earlier setup of 
> > solr3.6
> >
> 
> Just to confirm, are you using the backup command of replication
> handler? Can you post a thread dump of the process to pastebin or
> http://apaste.info/ ?
> 
> > Regards,
> >
> >
> >> Date: Fri, 5 Jul 2013 16:49:24 +0530
> >> Subject: Re: Solr 4.3 Master/Slave Issues
> >> From: shalinman...@gmail.com
> >> To: solr-user@lucene.apache.org
> >>
> >> This can mean multiple things:
> >> 1. You had killed a solr process earlier which left the lock file in place
> >> 2. You have more than one Solr core pointing to the same data directory
> >> 3. A solr process is already running and you are trying to start
> >> another one with the same config.
> >>
> >> On Fri, Jul 5, 2013 at 2:34 PM, Cool Techi  wrote:
> >> > We have set up solr 4.3 with master/setup and are facing a couple of 
> >> > issues,
> >> >
> >> > Index locking, the index on slave hangs at time and when we restart the 
> >> > core the core get's locked up. I have checked the logs and there are no 
> >> > OOM error or anything else other than the error given below,Caused by: 
> >> > org.apache.solr.common.SolrException: Index locked for write for core XXX
> >> > at org.apache.solr.core.SolrCore.(SolrCore.java:821)
> >> > at org.apache.solr.core.SolrCore.(SolrCore.java:618)
> >> > at 
> >> > org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:949)
> >> > at org.apache.solr.core.CoreContainer.create(CoreContainer.java:984)
> >> > at org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:597)
> >> > at org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:592)
> >> > at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
> >> > at java.util.concurrent.FutureTask.run(FutureTask.java:138)
> >> > at 
> >> > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
> >> > at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
> >> > at java.util.concurrent.FutureTask.run(FutureTask.java:138)
> >> > at 
> >> > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
> >> > at 
> >> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
> >> > ... 1 more
> >> > Caused by: org.apache.lucene.store.LockObtainFailedException: Index 
> >> > locked for write for core XXat 
> >> > org.apache.solr.core.SolrCore.initIndex(SolrCore.java:484)
> >> > at org.apache.solr.core.SolrCore.(SolrCore.java:745)
> >> > ... 13 moreWhen we fire backup on Slave, the whole core hangs after 
> >> > a while and also replication stops. This was not happening with our 
> >> > earlier setup of solr3.6
> >> >
> >> >
> >> >
> >> > Regards,
> >> > Ayush
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >>
> >>
> >>
> >> --
> >> Regards,
> >> Shalin Shekhar Mangar.
> >
> 
> 
> 
> --
> Regards,
> Shalin Shekhar Mangar.
  

Backup stops replication

2013-07-06 Thread Cool Techi
Hi,

We migrated from solr 3.6 to solr 4.3, when we fire a backup command 
(/replication?command=backup&location=/disk4/backups ) on the salve the slave 
stops replicating OR starts full replication from the master. This was not the 
behavior in the earlier version or solr, I have check the logs but dont see 
anything strange.

Correct me if I am wrong but isn't replication handler supposed to be async. I 
want the backup to be async how can this be achieved.

Regards,
Ayush
  

RE: Solr 4.3 Master/Slave Issues

2013-07-06 Thread Cool Techi
thanks Erick.

My tomcat setup is running on Ubuntu and has nothing else deployed other than 
the Solr war. My suspicion is it takes a long time to de-allocate the memory it 
has reserved for itself, but will get dump to find better.

regards,
Ayush

> Date: Sat, 6 Jul 2013 07:55:57 -0400
> Subject: Re: Solr 4.3 Master/Slave Issues
> From: erickerick...@gmail.com
> To: solr-user@lucene.apache.org
> 
> kill -9 is evil, you aren't certain what the state is afterwards,
> so the presence of the lock file is not surprising..
> 
> solrconfig.xml has a commented-out entry
> 
> false
> 
> I haven't personally used it, but it looks like it might
> help if you insist on kill -9.
> 
> Personally I'd try to figure out why your setup takes
> so long to shut down, I know lots of Tomcat installations
> that seem to function so this might point to something
> else in your setup that needs fixing.
> 
> Best
> Erick
> 
> 
> On Fri, Jul 5, 2013 at 11:54 AM, Cool Techi  wrote:
> 
> > The normal tomcat shutdown doesn't stop the server and take a long time,
> > so i do issue a kill -9 command. Any other suggestion to do this without
> > the locking.
> >
> > I would initiate a backup again and send the logs.
> >
> > regards,
> > Ayush
> >
> > > Date: Fri, 5 Jul 2013 19:40:12 +0530
> > > Subject: Re: Solr 4.3 Master/Slave Issues
> > > From: shalinman...@gmail.com
> > > To: solr-user@lucene.apache.org
> > >
> > > On Fri, Jul 5, 2013 at 6:14 PM, Cool Techi 
> > wrote:
> > > >
> > > > 1) That was my initial suspicion, but when I run ps -aux | grep
> > "java", but there it doesn't show any other program running. I kill the
> > process and start again and it locks.
> > >
> > > How are you killing the process? A SIGKILL will leave a lock file.
> > >
> > > >
> > > > 2) When we fire backup on Slave, the whole core hangs after a while
> > and also replication stops. This was not happening with our earlier setup
> > of solr3.6
> > > >
> > >
> > > Just to confirm, are you using the backup command of replication
> > > handler? Can you post a thread dump of the process to pastebin or
> > > http://apaste.info/ ?
> > >
> > > > Regards,
> > > >
> > > >
> > > >> Date: Fri, 5 Jul 2013 16:49:24 +0530
> > > >> Subject: Re: Solr 4.3 Master/Slave Issues
> > > >> From: shalinman...@gmail.com
> > > >> To: solr-user@lucene.apache.org
> > > >>
> > > >> This can mean multiple things:
> > > >> 1. You had killed a solr process earlier which left the lock file in
> > place
> > > >> 2. You have more than one Solr core pointing to the same data
> > directory
> > > >> 3. A solr process is already running and you are trying to start
> > > >> another one with the same config.
> > > >>
> > > >> On Fri, Jul 5, 2013 at 2:34 PM, Cool Techi 
> > wrote:
> > > >> > We have set up solr 4.3 with master/setup and are facing a couple
> > of issues,
> > > >> >
> > > >> > Index locking, the index on slave hangs at time and when we restart
> > the core the core get's locked up. I have checked the logs and there are no
> > OOM error or anything else other than the error given below,Caused by:
> > org.apache.solr.common.SolrException: Index locked for write for core XXX
> > > >> > at org.apache.solr.core.SolrCore.(SolrCore.java:821)
> > > >> > at org.apache.solr.core.SolrCore.(SolrCore.java:618)
> > > >> > at
> > org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:949)
> > > >> > at
> > org.apache.solr.core.CoreContainer.create(CoreContainer.java:984)
> > > >> > at
> > org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:597)
> > > >> > at
> > org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:592)
> > > >> > at
> > java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
> > > >> > at java.util.concurrent.FutureTask.run(FutureTask.java:138)
> > > >> > at
> > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
> &

solr cloud index corruption

2013-07-08 Thread Cool Techi
Hi,

We are frequently getting issues of index corruption on the cloud, this used to 
not happen in our master slave setup with solr 3.6. I have tried to check the 
logs, but don't see an exact reason.

I have run the index checker and it recovers, but I am not able to understand 
as to why this is happening. Any pointers would help.

regards,
rohit
  

Autowarming last 15 days data

2013-07-31 Thread Cool Techi
Hi,

We have a solr master slave set up with close to 30 million records. Our index 
changes/updates very frequently and replication is set up at 60 seconds delay.

Now every time replication completes, the new searches take a time. How can 
this be improved? I have come across that warming would help this scenario, I 
our case we cannot warm some queries, but most of the users use the last 15 
days data only. 

So would it be possible to auto warm only last 15 days data?

Regards,
Ayush
  

RE: Autowarming last 15 days data

2013-07-31 Thread Cool Techi
Would it make sense if we open a newSearcher with the last 15 days documents? 
since these is the documents which are mostly used by the users. Also, how 
could i do this if this is possible?


> Date: Wed, 31 Jul 2013 08:54:47 -0600
> From: s...@elyograg.org
> To: solr-user@lucene.apache.org
> Subject: Re: Autowarming last 15 days data
> 
> On 7/31/2013 7:30 AM, Cool Techi wrote:
> > We have a solr master slave set up with close to 30 million records. Our 
> > index changes/updates very frequently and replication is set up at 60 
> > seconds delay.
> > 
> > Now every time replication completes, the new searches take a time. How can 
> > this be improved? I have come across that warming would help this scenario, 
> > I our case we cannot warm some queries, but most of the users use the last 
> > 15 days data only. 
> > 
> > So would it be possible to auto warm only last 15 days data?
> 
> Autowarming is generally done automatically when a new searcher is
> opened, according to the cache config.  It will take the most recent N
> queries in the cache (according to the autowarmCount) and re-execute
> those queries against the index to populate the cache.  The document
> cache cannot be warmed directly, but when the query result cache is
> warmed, that will also populate the document cache.
> 
> Because you have a potentially very frequent interval for opening new
> searchers (possibly replicating every 60 seconds), you will want to
> avoid large autowarmCount values.  If your autowarming ends up taking
> too long, the system will try to open a new searcher while the previous
> one is being warmed, which can lead to problems.  I have found that the
> filterCache is particularly slow to warm.
> 
> Thanks,
> Shawn
> 
  

RE: Autowarming last 15 days data

2013-08-01 Thread Cool Techi
Thanks Shwan, get the point.

Regards,
Ayush

> Date: Wed, 31 Jul 2013 09:46:49 -0600
> From: s...@elyograg.org
> To: solr-user@lucene.apache.org
> Subject: Re: Autowarming last 15 days data
> 
> On 7/31/2013 9:21 AM, Cool Techi wrote:
> > Would it make sense if we open a newSearcher with the last 15 days 
> > documents? since these is the documents which are mostly used by the users. 
> > Also, how could i do this if this is possible?
> 
> When you open a searcher, it's for the entire index.  You may want to go
> distributed and keep the newest 15 days of data in a separate index from
> the rest.  For my own index, I use this hot/cold shard setup.  I have a
> nightly process that indexes data that needs to be moved into the cold
> shards and deletes it from the hot shard.
> 
> http://wiki.apache.org/solr/DistributedSearch
> 
> SolrCloud is the future of distributed search, but it does not have
> built-in support for a hot/cold shard setup.  You'd need to manage that
> yourself with manual sharding.  A custom sharding plugin to automate
> indexing would likely be very very involved, it would probably be easier
> to manage it outside of SolrCloud.
> 
> Thanks,
> Shawn
> 
  

Document Missing from Share in Solr cloud

2013-04-16 Thread Cool Techi
Hi,

We noticed a strange behavior in our solr cloud setup, we are using solr4.2  
with 1:3 replication setting. We noticed that some of the documents were 
showing up in search sometimes and not at other, the reason being the document 
was not present in all the shards.

We have restarted zookeeper and also entire cloud, but these documents are not 
being replicated in all the shards for some reason and hence inconsistent 
search results.

Regards,
Ayush
  

RE: Document Missing from Share in Solr cloud

2013-04-16 Thread Cool Techi
That's what I am trying to say, the document is not replicated across all the 
replicas for a specific shard, hence the query show different results on every 
refresh.



> Date: Tue, 16 Apr 2013 11:34:18 -0600
> Subject: Re: Document Missing from Share in Solr cloud
> From: thelabd...@gmail.com
> To: solr-user@lucene.apache.org
> 
> If you are using the default doc router for indexing in SolrCloud, then a
> document only exists in a single shard but can be replicated in that shard
> to any number of replicas.
> 
> Can you clarify your question as it sounds like you're saying that the
> document is not replicated across all the replicas for a specific shard? If
> so, that's definitely a problem ...
> 
> 
> On Tue, Apr 16, 2013 at 11:22 AM, Cool Techi  wrote:
> 
> > Hi,
> >
> > We noticed a strange behavior in our solr cloud setup, we are using
> > solr4.2  with 1:3 replication setting. We noticed that some of the
> > documents were showing up in search sometimes and not at other, the reason
> > being the document was not present in all the shards.
> >
> > We have restarted zookeeper and also entire cloud, but these documents are
> > not being replicated in all the shards for some reason and hence
> > inconsistent search results.
> >
> > Regards,
> > Ayush
> >
  

RE: Document Missing from Share in Solr cloud

2013-04-17 Thread Cool Techi
Field type is string and this has happened for multiple docs over the past week.

Regards,
Ayush

> Date: Tue, 16 Apr 2013 14:06:40 -0600
> Subject: Re: Document Missing from Share in Solr cloud
> From: thelabd...@gmail.com
> To: solr-user@lucene.apache.org
> 
> btw ... what is the field type of your unique ID field?
> 
> 
> On Tue, Apr 16, 2013 at 12:34 PM, Timothy Potter wrote:
> 
> > Ok, that makes more sense and is definitely cause for concern. Do you have
> > a sense for whether this is ongoing or happened a few times unexpectedly in
> > the past? If ongoing, then will probably be easier to track down the root
> > cause.
> >
> >
> > On Tue, Apr 16, 2013 at 12:08 PM, Cool Techi wrote:
> >
> >> That's what I am trying to say, the document is not replicated across all
> >> the replicas for a specific shard, hence the query show different results
> >> on every refresh.
> >>
> >>
> >>
> >> > Date: Tue, 16 Apr 2013 11:34:18 -0600
> >> > Subject: Re: Document Missing from Share in Solr cloud
> >> > From: thelabd...@gmail.com
> >> > To: solr-user@lucene.apache.org
> >> >
> >> > If you are using the default doc router for indexing in SolrCloud, then
> >> a
> >> > document only exists in a single shard but can be replicated in that
> >> shard
> >> > to any number of replicas.
> >> >
> >> > Can you clarify your question as it sounds like you're saying that the
> >> > document is not replicated across all the replicas for a specific
> >> shard? If
> >> > so, that's definitely a problem ...
> >> >
> >> >
> >> > On Tue, Apr 16, 2013 at 11:22 AM, Cool Techi 
> >> wrote:
> >> >
> >> > > Hi,
> >> > >
> >> > > We noticed a strange behavior in our solr cloud setup, we are using
> >> > > solr4.2  with 1:3 replication setting. We noticed that some of the
> >> > > documents were showing up in search sometimes and not at other, the
> >> reason
> >> > > being the document was not present in all the shards.
> >> > >
> >> > > We have restarted zookeeper and also entire cloud, but these
> >> documents are
> >> > > not being replicated in all the shards for some reason and hence
> >> > > inconsistent search results.
> >> > >
> >> > > Regards,
> >> > > Ayush
> >> > >
> >>
> >>
> >
> >
  

RE: Document Missing from Share in Solr cloud

2013-04-17 Thread Cool Techi
Shouldnt the number of docs across shards be same, I can see a difference 

Shard 1
Last Modified:about 2 hours agoNum Docs:26236135Max Doc:26592164Deleted 
Docs:356029Version:6672183Segment Count:34Shard1  Replica

Last Modified: about 2 hours agoNum Docs:26236135Max Doc:26594887Deleted 
Docs:358752Version:6678209Segment Count: 27

> From: cooltec...@outlook.com
> To: solr-user@lucene.apache.org
> Subject: RE: Document Missing from Share in Solr cloud
> Date: Wed, 17 Apr 2013 13:28:16 +0530
> 
> Field type is string and this has happened for multiple docs over the past 
> week.
> 
> Regards,
> Ayush
> 
> > Date: Tue, 16 Apr 2013 14:06:40 -0600
> > Subject: Re: Document Missing from Share in Solr cloud
> > From: thelabd...@gmail.com
> > To: solr-user@lucene.apache.org
> > 
> > btw ... what is the field type of your unique ID field?
> > 
> > 
> > On Tue, Apr 16, 2013 at 12:34 PM, Timothy Potter 
> > wrote:
> > 
> > > Ok, that makes more sense and is definitely cause for concern. Do you have
> > > a sense for whether this is ongoing or happened a few times unexpectedly 
> > > in
> > > the past? If ongoing, then will probably be easier to track down the root
> > > cause.
> > >
> > >
> > > On Tue, Apr 16, 2013 at 12:08 PM, Cool Techi 
> > > wrote:
> > >
> > >> That's what I am trying to say, the document is not replicated across all
> > >> the replicas for a specific shard, hence the query show different results
> > >> on every refresh.
> > >>
> > >>
> > >>
> > >> > Date: Tue, 16 Apr 2013 11:34:18 -0600
> > >> > Subject: Re: Document Missing from Share in Solr cloud
> > >> > From: thelabd...@gmail.com
> > >> > To: solr-user@lucene.apache.org
> > >> >
> > >> > If you are using the default doc router for indexing in SolrCloud, then
> > >> a
> > >> > document only exists in a single shard but can be replicated in that
> > >> shard
> > >> > to any number of replicas.
> > >> >
> > >> > Can you clarify your question as it sounds like you're saying that the
> > >> > document is not replicated across all the replicas for a specific
> > >> shard? If
> > >> > so, that's definitely a problem ...
> > >> >
> > >> >
> > >> > On Tue, Apr 16, 2013 at 11:22 AM, Cool Techi 
> > >> wrote:
> > >> >
> > >> > > Hi,
> > >> > >
> > >> > > We noticed a strange behavior in our solr cloud setup, we are using
> > >> > > solr4.2  with 1:3 replication setting. We noticed that some of the
> > >> > > documents were showing up in search sometimes and not at other, the
> > >> reason
> > >> > > being the document was not present in all the shards.
> > >> > >
> > >> > > We have restarted zookeeper and also entire cloud, but these
> > >> documents are
> > >> > > not being replicated in all the shards for some reason and hence
> > >> > > inconsistent search results.
> > >> > >
> > >> > > Regards,
> > >> > > Ayush
> > >> > >
> > >>
> > >>
> > >
> > >
> 
  

RE: Document Missing from Share in Solr cloud

2013-04-17 Thread Cool Techi
Sorry, made a copy paste mistake. The numbers are different.

My cloud has two shards with each shard having 1 replica. One of the shards and 
replica have the same number of docs, while in the other shard there is a 
mismatch.

Regards,
Ayush

> From: u...@odoko.co.uk
> To: solr-user@lucene.apache.org
> Subject: Re: Document Missing from Share in Solr cloud
> Date: Wed, 17 Apr 2013 09:48:03 +0100
> 
> Well, your numdocs *is* the same. Your maxdocs isn't, which sounds right
> to me.
> 
> maxdocs is the number of documents, including deleted ones. Given
> deleted docs are purged by background merges, it makes sense that each
> index is deciding differently when to do those merges. But the number of
> undeleted docs is the same which is a good thing.
> 
> Do queries against each replica for a shard, with distrib=false, and see
> whether the results are the same.
> 
> Upayavira
> 
> On Wed, Apr 17, 2013, at 09:14 AM, Cool Techi wrote:
> > Shouldnt the number of docs across shards be same, I can see a difference 
> > 
> > Shard 1
> > Last Modified:about 2 hours agoNum Docs:26236135Max Doc:26592164Deleted
> > Docs:356029Version:6672183Segment Count:34Shard1  Replica
> > 
> > Last Modified: about 2 hours agoNum Docs:26236135Max Doc:26594887Deleted
> > Docs:358752Version:6678209Segment Count: 27
> > 
> > > From: cooltec...@outlook.com
> > > To: solr-user@lucene.apache.org
> > > Subject: RE: Document Missing from Share in Solr cloud
> > > Date: Wed, 17 Apr 2013 13:28:16 +0530
> > > 
> > > Field type is string and this has happened for multiple docs over the 
> > > past week.
> > > 
> > > Regards,
> > > Ayush
> > > 
> > > > Date: Tue, 16 Apr 2013 14:06:40 -0600
> > > > Subject: Re: Document Missing from Share in Solr cloud
> > > > From: thelabd...@gmail.com
> > > > To: solr-user@lucene.apache.org
> > > > 
> > > > btw ... what is the field type of your unique ID field?
> > > > 
> > > > 
> > > > On Tue, Apr 16, 2013 at 12:34 PM, Timothy Potter 
> > > > wrote:
> > > > 
> > > > > Ok, that makes more sense and is definitely cause for concern. Do you 
> > > > > have
> > > > > a sense for whether this is ongoing or happened a few times 
> > > > > unexpectedly in
> > > > > the past? If ongoing, then will probably be easier to track down the 
> > > > > root
> > > > > cause.
> > > > >
> > > > >
> > > > > On Tue, Apr 16, 2013 at 12:08 PM, Cool Techi 
> > > > > wrote:
> > > > >
> > > > >> That's what I am trying to say, the document is not replicated 
> > > > >> across all
> > > > >> the replicas for a specific shard, hence the query show different 
> > > > >> results
> > > > >> on every refresh.
> > > > >>
> > > > >>
> > > > >>
> > > > >> > Date: Tue, 16 Apr 2013 11:34:18 -0600
> > > > >> > Subject: Re: Document Missing from Share in Solr cloud
> > > > >> > From: thelabd...@gmail.com
> > > > >> > To: solr-user@lucene.apache.org
> > > > >> >
> > > > >> > If you are using the default doc router for indexing in SolrCloud, 
> > > > >> > then
> > > > >> a
> > > > >> > document only exists in a single shard but can be replicated in 
> > > > >> > that
> > > > >> shard
> > > > >> > to any number of replicas.
> > > > >> >
> > > > >> > Can you clarify your question as it sounds like you're saying that 
> > > > >> > the
> > > > >> > document is not replicated across all the replicas for a specific
> > > > >> shard? If
> > > > >> > so, that's definitely a problem ...
> > > > >> >
> > > > >> >
> > > > >> > On Tue, Apr 16, 2013 at 11:22 AM, Cool Techi 
> > > > >> > 
> > > > >> wrote:
> > > > >> >
> > > > >> > > Hi,
> > > > >> > >
> > > > >> > > We noticed a strange behavior in our solr cloud setup, we are 
> > > > >> > > using
> > > > >> > > solr4.2  with 1:3 replication setting. We noticed that some of 
> > > > >> > > the
> > > > >> > > documents were showing up in search sometimes and not at other, 
> > > > >> > > the
> > > > >> reason
> > > > >> > > being the document was not present in all the shards.
> > > > >> > >
> > > > >> > > We have restarted zookeeper and also entire cloud, but these
> > > > >> documents are
> > > > >> > > not being replicated in all the shards for some reason and hence
> > > > >> > > inconsistent search results.
> > > > >> > >
> > > > >> > > Regards,
> > > > >> > > Ayush
> > > > >> > >
> > > > >>
> > > > >>
> > > > >
> > > > >
> > > 
> >   
  

Solr Cloud OR Load Balancing

2013-04-20 Thread Cool Techi
We are currently using Solr 3.6 with 1:1 replication on our production servers, 
but now out data has increased a lot and were looking to migrate to something 
more scalable. 

We have been experimenting with Solr Cloud over the past 2 months and noticed 
servers issues, most of them have been fixed in the version releases but some 
am not sure of, like
Docs missing in replica's of a shard.
Giving this what would the recommendation be, should we use solr4.2 with 
replication and load balancing using HA proxy so that we dont run into issue OR 
do we think Solr Cloud is good for live usage.
Regards,
Ayush
  

Solr Hangs on startup

2013-06-14 Thread Cool Techi
Hi,

We are using solr4.3 cloud setup, but for some reason solr fails to startup, I 
see the following in the log file, post this there are no logs.,

org.apache.solr.search.SolrIndexSearcher  ? Opening Searcher@17586ed7 main
15293 [coreLoadExecutor-3-thread-1] INFO  org.apache.solr.core.SolrCore  ? 
[cmn_shard1_replica1] Registered new searcher Searcher@17586ed7 
main{StandardDirectoryReader(segments_zpt:1650158 _6nln(4.3):C2734441/433094 
_6rvw(4.3):C3395530/166626 _6vdg(4.3):C4039667/172929 _6z52(4.3):C4137543/2279 
_770z(4.3):C5879498/97346 _71rz(4.3):C4168660/440273 _74hn(4.3):C5900928/134106 
_7a5j(4.3):C5892645/269769 _7bcs(4.3):C5502048/2562430 
_7gnp(4.3):C2310243/676016 _7klj(4.3):C3753172/184399 _7nxn(4.3):C558455


A thread dump show the following,

"coreLoadExecutor-3-thread-1" prio=10 tid=0x40b44800 nid=0x27ad 
runnable [0x7fbbff2d6000]
   java.lang.Thread.State: RUNNABLE
at sun.nio.ch.FileDispatcher.pread0(Native Method)
at sun.nio.ch.FileDispatcher.pread(FileDispatcher.java:31)
at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:199)
at sun.nio.ch.IOUtil.read(IOUtil.java:175)
at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:612)
at 
org.apache.solr.update.ChannelFastInputStream.readWrappedStream(TransactionLog.java:752)
at 
org.apache.solr.common.util.FastInputStream.refill(FastInputStream.java:89)
at 
org.apache.solr.common.util.FastInputStream.peek(FastInputStream.java:61)
at 
org.apache.solr.update.TransactionLog$ReverseReader.next(TransactionLog.java:702)
at 
org.apache.solr.update.UpdateLog$RecentUpdates.update(UpdateLog.java:925)
at 
org.apache.solr.update.UpdateLog$RecentUpdates.access$000(UpdateLog.java:863)
at 
org.apache.solr.update.UpdateLog.getRecentUpdates(UpdateLog.java:1014)
at org.apache.solr.update.UpdateLog.init(UpdateLog.java:253)
at org.apache.solr.update.UpdateHandler.initLog(UpdateHandler.java:82)
at org.apache.solr.update.UpdateHandler.(UpdateHandler.java:137)
at org.apache.solr.update.UpdateHandler.(UpdateHandler.java:123)
at 
org.apache.solr.update.DirectUpdateHandler2.(DirectUpdateHandler2.java:95)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
at org.apache.solr.core.SolrCore.createInstance(SolrCore.java:525)
at org.apache.solr.core.SolrCore.createUpdateHandler(SolrCore.java:596)
at org.apache.solr.core.SolrCore.(SolrCore.java:805)
at org.apache.solr.core.SolrCore.(SolrCore.java:618)
at 
org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:949)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:984)
at org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:597)
at org.apache.solr.core.CoreContainer$2.call(CoreContainer.java:592)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)


What could be causing this?

regards,
Ayush
  

Single Tomcat Multiple Shards

2012-11-21 Thread Cool Techi
Hey Guys,

We are experimenting with solr cloud, this is what we want to set up as,

2 Machines each having have 8 master shards, so total of 16 shards. The 
assumption is we want to store approximately 4-5 TB data over a period of 1 
year of so. 
Replication factor of 1 which are again distributed across 3-4 machines.
Initially we want to start with 8 shards in a single tomcat and single machine, 
but I cannot find a way of having multiple shards in a single SOLR_HOME and 
single Tomcat. Can this be achieved?
Regards,Ayush

  

Reloading config to zookeeper

2012-11-22 Thread Cool Techi
When we make changes to our config files, how do we reload the files into 
zookeeper. 

Also, I understand that we would need to reload the collection, would we need 
to do this at a per shard level or just at the cloud level.

Regards,
Ayush

  

RE: Reloading config to zookeeper

2012-11-22 Thread Cool Techi
Thanks, but why do we need to specify the -solrhome? 

I am using the following command to load new config,

java -classpath .:/Users/solr-cli-lib/* org.apache.solr.cloud.ZkCLI -cmd 
upconfig -zkhost 
localhost:2181,localhost:2182,localhost:2183,localhost:2184,localhost:2185 
-confdir /Users/config-files -confname myconf

So basically reloading is just uploading the configs back again?

Regard,s
Ayush

> Date: Thu, 22 Nov 2012 19:32:27 +0100
> Subject: Re: Reloading config to zookeeper
> From: mrzewu...@gmail.com
> To: solr-user@lucene.apache.org
> 
> Hi,
> 
> I'm using "cloud-scripts/zkcli.sh" script for reloading configuration, for
> example:
> $ ./cloud-scripts/zkcli.sh -cmd upconfig -confdir  -solrhome
>  -confname  -z 
> 
> Then I'm reloading collection on each node in cloud, but maybe someone
> knows better solution.
> Regards.
> 
> On 22 November 2012 19:23, Cool Techi  wrote:
> 
> > When we make changes to our config files, how do we reload the files into
> > zookeeper.
> >
> > Also, I understand that we would need to reload the collection, would we
> > need to do this at a per shard level or just at the cloud level.
> >
> > Regards,
> > Ayush
> >
> >
  

Solr Cloud 4.0 Production Ready?

2012-12-17 Thread Cool Techi
Hi,

We have been using solr 3.5 in our production for sometime now and facing the 
problems faced by a large solr index. We wanted to migrate to Solr Cloud and 
have started some experimentation. But in the mean time also following the user 
forum and seem to be noticing a lot of bugs which were raised post the release 
and will be fixed in 4.1.

Should we wait for 4.1 release for production or we can go ahead with the 
current release?

Regards,
Ayush


  

Searching for Solr Stop Words

2013-01-05 Thread Cool Techi

On of my solr fields is configured in the following manned,


 





 
 
   
   
   
   




This works in cases where i don't want stemming, but now there is 
another use case which is causing a problem, people are beginning to 
seach for the following combinations,

The Ivy : In this case results with just ivy is being returned, 
when the expected result would be with The. I understand that this is 
because of the stop word but is the way to achieve this. For example if 
they search for "the ivy" within quotes than this should work.(Mom & Me) OR 
("mom and me"): In this case also & is 
dropped or results including both mom and me in some part of the 
statement is returned.

I am ok if only new data behaves in the right way but wouldnt be able
 to reindex. Also, would changing the schema.xml file trigger a full 
replication?


Regards,

Ayush

  

Solr commit taking too long

2013-01-16 Thread Cool Techi
Hi,

We have an index of approximately 400GB in size, indexing 5000 documents was 
taking 20 seconds. But lately, the indexing is taking very long, committing the 
same amount of document is taking 5-20 mins. 

On checking the logs I can see that their a frequent merges happening, which I 
am guessing is the reason for this, how can this be improved. My configurations 
are given below,

false
30
64

regards,
Ayush
  

Not condition not working for Korean search

2013-02-07 Thread Cool Techi
Hi,

I am no Korean expert and am finding it difficult to fix this, my client is 
searching for the following query, but the NOT condition doesn't seem to be 
working.

(stnostem:((옵티머스 OR "엘지 스마트폰") AND NOT ("옵티머스 프라임" OR 프라임)))

 the search result (xml attached ) return result with the not condition 
keywords? How can this be fixed.

Regards,
Ayush
  

Hourly Faceting

2013-02-08 Thread Cool Techi
Hi,

I want to facet results on an hourly basis, the following query gives me an 
hourly breakdown, but with the date part. I want just the hour part across the 
days. Is there any other way of doing this,


2013-02-01T00:00:00Z-330MINUTES
true
twitterId:191343557
createdOnGMTDate
+1HOUR
2013-02-08T23:59:59Z-330MINUTES
0


Result

0
0
0
0
0
0

Desired Result

0
0
0
0
0
0

Regards,
Ayush
  

Not condition not working for Korean search

2013-02-10 Thread Cool Techi
Any help here, still not able to understand what might be causing this. I have 
tried all combinations below, but to no avail,

(stnostem:((옵티머스 OR "엘지 스마트폰") AND NOT stnostem:("옵티머스 프라임" OR 프라임)))

(stnostem:((옵티머스 OR "엘지 스마트폰") AND - stnostem:("옵티머스 프라임" OR 프라임)))

What could be wrong here.



From: cooltec...@outlook.com
To: solr-user@lucene.apache.org
Subject: Not condition not working for Korean search
Date: Thu, 7 Feb 2013 19:46:13 +0530




Hi,

I am no Korean expert and am finding it difficult to fix this, my client is 
searching for the following query, but the NOT condition doesn't seem to be 
working.

(stnostem:((옵티머스 OR "엘지 스마트폰") AND NOT ("옵티머스 프라임" OR 프라임)))

 the search result (xml attached ) return result with the not condition 
keywords? How can this be fixed.

Regards,
Ayush

  

RE: Not condition not working for Korean search

2013-02-10 Thread Cool Techi
 idf(docFreq=5632, maxDocs=290963725)
0.034640186 = queryNorm
7.407695 = (MATCH) fieldWeight(stnostem:옵티머스 in 2006207), product of:
1.0 = tf(termFreq(stnostem:옵티머스)=1)
11.852312 = idf(docFreq=5632, maxDocs=290963725)
0.625 = fieldNorm(field=stnostem, doc=2006207)
0.5 = coord(1/2)

1.520675 = (MATCH) sum of:
1.520675 = (MATCH) product of:
3.04135 = (MATCH) sum of:
3.04135 = (MATCH) weight(stnostem:옵티머스 in 1338974), product of:
0.4105663 = queryWeight(stnostem:옵티머스), product of:
11.852312 = idf(docFreq=5632, maxDocs=290963725)
0.034640186 = queryNorm
7.407695 = (MATCH) fieldWeight(stnostem:옵티머스 in 1338974), product of:
1.0 = tf(termFreq(stnostem:옵티머스)=1)
11.852312 = idf(docFreq=5632, maxDocs=290963725)
0.625 = fieldNorm(field=stnostem, doc=1338974)
0.5 = coord(1/2)




> From: j...@basetechnology.com
> To: solr-user@lucene.apache.org
> Subject: Re: Not condition not working for Korean search
> Date: Mon, 11 Feb 2013 00:37:03 -0500
> 
> Show us the parsed query that you get with &debugQuery=true
> 
> Actually this may be the "pure negative" bug where you need to change:
> 
> A AND -B
> 
> into
> 
> A AND (*:* -B)
> 
> -- Jack Krupansky
> 
> -Original Message- 
> From: Cool Techi
> Sent: Monday, February 11, 2013 12:30 AM
> To: solr-user@lucene.apache.org
> Subject: Not condition not working for Korean search
> 
> Any help here, still not able to understand what might be causing this. I 
> have tried all combinations below, but to no avail,
> 
> (stnostem:((옵티머스 OR "엘지 스마트폰") AND NOT stnostem:("옵티머스 프라임" 
> OR 프라임)))
> 
> (stnostem:((옵티머스 OR "엘지 스마트폰") AND - stnostem:("옵티머스 프라임" 
> OR 프라임)))
> 
> What could be wrong here.
> 
> 
> 
> From: cooltec...@outlook.com
> To: solr-user@lucene.apache.org
> Subject: Not condition not working for Korean search
> Date: Thu, 7 Feb 2013 19:46:13 +0530
> 
> 
> 
> 
> Hi,
> 
> I am no Korean expert and am finding it difficult to fix this, my client is 
> searching for the following query, but the NOT condition doesn't seem to be 
> working.
> 
> (stnostem:((옵티머스 OR "엘지 스마트폰") AND NOT ("옵티머스 프라임" OR 
> 프라임)))
> 
> the search result (xml attached ) return result with the not condition 
> keywords? How can this be fixed.
> 
> Regards,
> Ayush
>  
> 
  

SEVERE RecoveryStrategy Recovery failed - trying again... (9)

2013-02-18 Thread Cool Techi
I am seeing the following error in my Admin console and the core/ cloud status 
is taking forever to load.

SEVERERecoveryStrategyRecovery failed - trying again... (9) 

What causes this and how can I recover from this mode?

Regards,
Rohit

  

RE: SEVERE RecoveryStrategy Recovery failed - trying again... (9)

2013-02-18 Thread Cool Techi
There is not error I can see in the logs, my shards are divided over three 
machines, the cloud runs fine when I don't bring up one of the nodes, the 
moment I start that particular note, the cloud stops responding,

Feb 19, 2013 5:22:22 AM 
org.apache.solr.handler.component.SpellCheckComponent$SpellCheckerListener 
newSearcher
INFO: Loading spell index for spellchecker: default
Feb 19, 2013 5:22:22 AM 
org.apache.solr.handler.component.SpellCheckComponent$SpellCheckerListener 
newSearcher
INFO: Loading spell index for spellchecker: wordbreak
Feb 19, 2013 5:22:22 AM org.apache.solr.core.SolrCore registerSearcher
INFO: [cmn] Registered new searcher Searcher@3b47788d 
main{StandardDirectoryReader(segments_1dvf:1488121 _2acm(4.1):C13967428/87404 
_62w6(4.1):C259989/31792 _8ehw(4.1):C405062/57136 _8um4(4.1):C228434/26526 
_a0i1(4.1):C171825/43653 _bgu3(4.1):C315311/30246 _ao6h(4.1):C176468/44702 
_b7uu(4.1):C97823/27124 _bjzb(4.1):C77280/8476 _bra3(4.1):C142681/21340 
_bzpo(4.1):C198058/23506 _c0jh(4.1):C18201/8171 _c307(4.1):C37984/5305 
_c2e0(4.1):C22300/9788 _c1o6(4.1):C23523/8630 _c3hl(4.1):C12034/2871 
_c3kw(4.1):C5821/971 _c3l6(4.1):C1106 _c3lh(4.1):C707/1 _c3lu(4.1):C509/2 
_c3mf(4.1):C482/1 _c3m5(4.1):C374/2 _c3mc(4.1):C164/2 _c3mh(4.1):C64/3 
_c3mi(4.1):C49 _c3mj(4.1):C25 _c3mk(4.1):C12)}
Feb 19, 2013 5:22:22 AM org.apache.solr.cloud.ZkController publish
INFO: publishing core=cmn state=down
Feb 19, 2013 5:22:22 AM org.apache.solr.cloud.ZkController publish
INFO: numShards not found on descriptor - reading it from system property
Feb 19, 2013 5:22:22 AM org.apache.solr.core.CoreContainer registerCore
INFO: registering core: cmn
Feb 19, 2013 5:22:22 AM org.apache.solr.cloud.ZkController register
INFO: Register replica - core:cmn address:http://10.0.0.205:8080/solr 
collection:cmn shard:shard2
Feb 19, 2013 5:22:22 AM org.apache.solr.client.solrj.impl.HttpClientUtil 
createClient
INFO: Creating new http client, 
config:maxConnections=1&maxConnectionsPerHost=20&connTimeout=3&socketTimeout=3&retry=false


Regards,
Ayush

> Subject: Re: SEVERERecoveryStrategyRecovery failed - trying again... 
> (9)
> From: markrmil...@gmail.com
> Date: Mon, 18 Feb 2013 10:21:53 -0500
> To: solr-user@lucene.apache.org
> 
> We need to see more of your logs to determine why - there should be some 
> exceptions logged.
> 
> - Mark
> 
> On Feb 18, 2013, at 9:47 AM, Cool Techi  wrote:
> 
> > I am seeing the following error in my Admin console and the core/ cloud 
> > status is taking forever to load.
> > 
> > SEVERERecoveryStrategyRecovery failed - trying again... (9) 
> > 
> > What causes this and how can I recover from this mode?
> > 
> > Regards,
> > Rohit
> > 
> >   
> 
  

Solr on https

2012-09-11 Thread Cool Techi
Hi,

We are trying to run solr on https, these are few of the issues or problems 
that are coming up. Just wanted to understand if anyone else is facing these 
problems,

we have some shards running on https, but in shards parameter in solr we don't 
specify the protocol, how can we achieve thisWill replication work on https
Will commit and other functions working normally?
Regards,atpug
  

Solr Sharding Or Solr Cloud

2012-10-09 Thread Cool Techi
Hi,

We have reached a state where a single solr server is not proving to be enough 
and are exploring options of sharding or solr cloud. 

Just wanted to check the pros and cons of each one of them if there are any. 
What would be the best size of the shards or 

Regards,
R
  

New data not indexing, after data deletion.

2012-10-18 Thread Cool Techi
Hi,

We have a huge index size, running to around 350Gb+, we yesterday deleted some 
data on it. The deletion completed successfully, but post the deletion new data 
was not being indexed. We restarted the server, but neither the tomcat logs or 
Solr logs are giving any indication of why this might be happening.  All we 
keep getting is a timeout exception, but at the same time all the queries are 
running fine and returning normal results.

Has anyone faced a similar issue earlier or if anyone can point to how we debug 
this?

Regards
  

Architecture Question

2012-11-16 Thread Cool Techi



Hi,

I am not sure if this is the right forum for this question, but it would be 
great if I could be pointed in the right direction. We have been using a 
combination of MySql and Solr for all our company full text and query needs.  
But as our customers have grow so has the amount of data and MySql is just not 
proving to be a right option for storing/querying.

I have been looking at Solr Cloud and it looks really impressive, but and not 
sure if we should give away our storage system. So, I have been exploring 
DataStax but a commercial option is out of question. So we were thinking of 
using hbase to store the data and at the same time index the data into Solr 
cloud, but for many reasons this design doesn't seem convincing (Also seen 
basic of Lilly).

1) Would it be recommended to just user Solr cloud with multiple replication or 
hbase-solr seems like good option
2) How much strain would be to keep both Solr Shard and Hbase node on the same 
machine
3) if there a calculation on what kind of machine configuration would I need to 
store 500-1000 million records. Most of these with be social data 
(Twitter/facebook/blogs etc) and how many shards.

Regards,
Ayush 


  

RE: Architecture Question

2012-11-16 Thread Cool Techi
Hi Otis,

Thanks for your reply, just wanted to check what NoSql structure would be best 
suited to store data and use the least amount of memory, since for most of my 
work Solr would be sufficient and I want to store data just in case we want to 
reindex and as a backup.

Regards,
Ayush

> Date: Fri, 16 Nov 2012 15:47:40 -0500
> Subject: Re: Architecture Question
> From: otis.gospodne...@gmail.com
> To: solr-user@lucene.apache.org
> 
> Hello,
> 
> 
> 
> > I am not sure if this is the right forum for this question, but it would
> > be great if I could be pointed in the right direction. We have been using a
> > combination of MySql and Solr for all our company full text and query
> > needs.  But as our customers have grow so has the amount of data and MySql
> > is just not proving to be a right option for storing/querying.
> >
> > I have been looking at Solr Cloud and it looks really impressive, but and
> > not sure if we should give away our storage system. So, I have been
> > exploring DataStax but a commercial option is out of question. So we were
> > thinking of using hbase to store the data and at the same time index the
> > data into Solr cloud, but for many reasons this design doesn't seem
> > convincing (Also seen basic of Lilly).
> >
> > 1) Would it be recommended to just user Solr cloud with multiple
> > replication or hbase-solr seems like good option
> >
> 
> If you trust SolrCloud with replication and keep all your fields stored
> then you could live without an external DB.  At this point I personally
> would still want an external DB.  Whether HBase is the right DB for the job
> I can't tell because I don't know anything about your data, volume, access
> patterns, etc.  I can tell you that HBase does scale well - we have tables
> with many billions of rows stored in it for instance.
> 
> 
> > 2) How much strain would be to keep both Solr Shard and Hbase node on the
> > same machine
> >
> 
> HBase loves memory.  So does Solr.  They both dislike disk IO (who
> doesn't!).  Solr can use a lot of CPU for indexing/searching, depending on
> the volume.  HBase RegionServers can use a lot of CPU if you run MapReuce
> on data in HBase.
> 
> 
> > 3) if there a calculation on what kind of machine configuration would I
> > need to store 500-1000 million records. Most of these with be social data
> > (Twitter/facebook/blogs etc) and how many shards.
> >
> 
> No recipe here, unfortunately.  You'd have to experiment and test, do load
> and performance testing, etc.  If you need help with Solr + HBase, we
> happen to have a lot of experience with both and have even used them
> together for some of our clients.
> 
> Otis
> --
> Performance Monitoring - http://sematext.com/spm/index.html
> Search Analytics - http://sematext.com/search-analytics/index.html
  

Question about Solr Cloud

2012-11-16 Thread Cool Techi
Hi,

I have just started working with Solr cloud and have a few questions related to 
the same,

1) In the start script we provide the the following, what's the purpose of 
providing this.
-Dbootstrap_confdir=./solr/collection1/conf Since we don't yet have a config in 
zookeeper, this parameter causes the local configuration directory ./solr/conf 
to be uploaded as the "myconf" config.  The name "myconf" is taken from the 
"collection.configName" param below.   -Dcollection.configName=myconf sets the 
config to use for the new collection. Omitting this param will cause the config 
name to default to "configuration1" 
2) When we make any changes into the config/schema do we need to copy it to all 
the shards running in the cloud manually?3) If we want to start with 10 shards 
on 2 machines, anticipating the future growth, do all these shards needs to run 
on separate jetty instances4) Is there any advantage of running solr on jetty 
then Tomcat?
Thanks,Ayush