grouping in solr cloud shard replicas

2018-01-06 Thread SANJAY.
Hi,

Please let me know how to achieve group by in solr could env.
We tried grouping in solr cloud shard replicas to fetch unique search result
from solr for custom field we .

we are getting exception saying unexpected docvalues type "SORTED_SET
(expected SORTED)"

We are using solr cloud and collection having 2 shared replicas .We have
created custom field type which is using solr.TextField class .

Please suggest me the best possible way to fetch the unique search result. 



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Solr cloud replica nodes missing some documents

2017-08-18 Thread Sanjay Lokhande
java:58)
at org.apache.solr.handler.UpdateRequestHandler$1.load
(UpdateRequestHandler.java:98)
at
org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody
(ContentStreamHandlerBase.java:74)
at org.apache.solr.handler.RequestHandlerBase.handleRequest
(RequestHandlerBase.java:143)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:2064)
at org.apache.solr.servlet.HttpSolrCall.execute
(HttpSolrCall.java:654)
at org.apache.solr.servlet.HttpSolrCall.call
(HttpSolrCall.java:450)
at org.apache.solr.servlet.SolrDispatchFilter.doFilter
(SolrDispatchFilter.java:227)
at org.apache.solr.servlet.SolrDispatchFilter.doFilter
(SolrDispatchFilter.java:196)
at org.eclipse.jetty.servlet.ServletHandler
$CachedChain.doFilter(ServletHandler.java:1652)
at org.eclipse.jetty.servlet.ServletHandler.doHandle
(ServletHandler.java:585)
at org.eclipse.jetty.server.handler.ScopedHandler.handle
(ScopedHandler.java:143)
at org.eclipse.jetty.security.SecurityHandler.handle
(SecurityHandler.java:577)
at org.eclipse.jetty.server.session.SessionHandler.doHandle
(SessionHandler.java:223)
at org.eclipse.jetty.server.handler.ContextHandler.doHandle
(ContextHandler.java:1127)
at org.eclipse.jetty.servlet.ServletHandler.doScope
(ServletHandler.java:515)
at org.eclipse.jetty.server.session.SessionHandler.doScope
(SessionHandler.java:185)
at org.eclipse.jetty.server.handler.ContextHandler.doScope
(ContextHandler.java:1061)
at org.eclipse.jetty.server.handler.ScopedHandler.handle
(ScopedHandler.java:141)
at
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle
(ContextHandlerCollection.java:215)
at org.eclipse.jetty.server.handler.HandlerCollection.handle
(HandlerCollection.java:110)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle
(HandlerWrapper.java:97)
at org.eclipse.jetty.server.Server.handle(Server.java:497)
at org.eclipse.jetty.server.HttpChannel.handle
(HttpChannel.java:310)
at org.eclipse.jetty.server.HttpConnection.onFillable
(HttpConnection.java:257)
at org.eclipse.jetty.io.AbstractConnection$2.run
(AbstractConnection.java:540)
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob
(QueuedThreadPool.java:635)
at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run
(QueuedThreadPool.java:555)
at java.lang.Thread.run(Thread.java:785)
Caused by: java.util.concurrent.TimeoutException: Idle timeout
expired: 5/5 ms
at org.eclipse.jetty.io.IdleTimeout.checkIdleTimeout
(IdleTimeout.java:161)
at org.eclipse.jetty.io.IdleTimeout$1.run(IdleTimeout.java:50)
at java.util.concurrent.Executors$RunnableAdapter.call
(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:267)
at java.util.concurrent.ScheduledThreadPoolExecutor
$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor
$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker
(ThreadPoolExecutor.java:1143)
at java.util.concurrent.ThreadPoolExecutor$Worker.run
(ThreadPoolExecutor.java:618)
... 1 more

Thanks in advance.

Regards,
Sanjay



Re: Solr cloud replica nodes missing some documents

2017-08-21 Thread Sanjay Lokhande


Any pointer guys.



From:   Sanjay Lokhande/India/IBM
To: solr-user@lucene.apache.org
Date:   08/18/2017 02:12 PM
Subject:Solr cloud replica nodes missing some documents


Hello  guys,

  I am having 5 nodes solr cloud setup with single shard. The solr version
is 5.2.1.
  server1 (http://146.XXX.com:4001/solr/contracts_shard1_replica4)is the
leader.
  A document with id '43e14a86cbdd422880cac22d9a15d3c0' was not replicated
3 nodes.
  Log shows that the "{add=[43e14a86cbdd422880cac22d9a15d3c0
(1573510697298427904)]}" request is received only by leader and server5
node.
  The server2, server3 & server4 node did not receive the request and hence
the document is missing in these nodes.

 Search "43e14a86cbdd422880cac22d9a15d3c0 "
  C:\solrIssue\solr_server1.log
INFO  - 2017-07-21 05:54:59.853; [contracts shard1 core_node2
contracts_shard1_replica4]
org.apache.solr.update.processor.LogUpdateProcessor;
[contracts_shard1_replica4] webapp=/solr path=/update params=
{wt=javabin&version=2} {deleteByQuery=id:(9467353f398448788c261aa347d75b8b
93332ab7f7ff4141a371713871ab65ad 8568e0eab8364bfc89c876aadfa01022
43e14a86cbdd422880cac22d9a15d3c0 a0af8cb24ef94d25b9691eee1f7024ca 8ad...
INFO  - 2017-07-21 05:54:59.853; [contracts shard1 core_node2
contracts_shard1_replica4]
org.apache.solr.update.processor.LogUpdateProcessor;
[contracts_shard1_replica4] webapp=/solr path=/update params=
{wt=javabin&version=2} {deleteByQuery=id:(9467353f398448788c261aa347d75b8b
93332ab7f7ff4141a371713871ab65ad 8568e0eab8364bfc89c876aadfa01022
43e14a86cbdd422880cac22d9a15d3c0 a0af8cb24ef94d25b9691eee1f7024ca 8ad...
INFO  - 2017-07-21 05:59:23.845; [contracts shard1 core_node2
contracts_shard1_replica4]
org.apache.solr.update.processor.LogUpdateProcessor;
[contracts_shard1_replica4] webapp=/solr path=/update params=
{wt=javabin&version=2} {add=[43e14a86cbdd422880cac22d9a15d3c0
(1573510697298427904)]} 0 26582
  C:\solrIssue\solr_server2\solr.log.1
INFO  - 2017-07-21 05:54:59.595; [contracts shard1 core_node4
contracts_shard1_replica5]
org.apache.solr.update.processor.LogUpdateProcessor;
[contracts_shard1_replica5] webapp=/solr path=/update params=
{update.distrib=FROMLEADER&_version_=-1573510446380482560&distrib.from=http://146.XXX.com:4001/solr/contracts_shard1_replica4/&wt=javabin&version=2}
 {deleteByQuery=id:(9467353f398448788c261aa347d75b8b
93332ab7f7ff4141a371713871ab65ad 8568e0eab8364bfc89c876aadfa01022
43e14a86cbdd422880cac22d9a15d3c0 a0af8cb24ef94d25b9691eee1f7024ca 8ad...
INFO  - 2017-07-21 05:54:59.595; [contracts shard1 core_node4
contracts_shard1_replica5]
org.apache.solr.update.processor.LogUpdateProcessor;
[contracts_shard1_replica5] webapp=/solr path=/update params=
{update.distrib=FROMLEADER&_version_=-1573510446380482560&distrib.from=http://146.XXX.com:4001/solr/contracts_shard1_replica4/&wt=javabin&version=2}
 {deleteByQuery=id:(9467353f398448788c261aa347d75b8b
93332ab7f7ff4141a371713871ab65ad 8568e0eab8364bfc89c876aadfa01022
43e14a86cbdd422880cac22d9a15d3c0 a0af8cb24ef94d25b9691eee1f7024ca 8ad...
  C:\solrIssue\solr_server3.log
INFO  - 2017-07-21 05:54:59.844; [contracts shard1 core_node1
contracts_shard1_replica3]
org.apache.solr.update.processor.LogUpdateProcessor;
[contracts_shard1_replica3] webapp=/solr path=/update params=
{update.distrib=FROMLEADER&_version_=-1573510446380482560&distrib.from=http://146.XXX.com:4001/solr/contracts_shard1_replica4/&wt=javabin&version=2}
 {deleteByQuery=id:(9467353f398448788c261aa347d75b8b
93332ab7f7ff4141a371713871ab65ad 8568e0eab8364bfc89c876aadfa01022
43e14a86cbdd422880cac22d9a15d3c0 a0af8cb24ef94d25b9691eee1f7024ca 8ad...
INFO  - 2017-07-21 05:54:59.844; [contracts shard1 core_node1
contracts_shard1_replica3]
org.apache.solr.update.processor.LogUpdateProcessor;
[contracts_shard1_replica3] webapp=/solr path=/update params=
{update.distrib=FROMLEADER&_version_=-1573510446380482560&distrib.from=http://146.XXX.com:4001/solr/contracts_shard1_replica4/&wt=javabin&version=2}
 {deleteByQuery=id:(9467353f398448788c261aa347d75b8b
93332ab7f7ff4141a371713871ab65ad 8568e0eab8364bfc89c876aadfa01022
43e14a86cbdd422880cac22d9a15d3c0 a0af8cb24ef94d25b9691eee1f7024ca 8ad...
  C:\solrIssue\solr_server4\solr.log.1
INFO  - 2017-07-21 05:54:59.734; [contracts shard1 core_node3
contracts_shard1_replica1]
org.apache.solr.update.processor.LogUpdateProcessor;
[contracts_shard1_replica1] webapp=/solr path=/update params=
{update.distrib=FROMLEADER&_version_=-1573510446380482560&distrib.from=http://146.XXX.com:4001/solr/contracts_shard1_replica4/&wt=javabin&version=2}
 {deleteByQuery=id:(9467353f398448788c261aa347d75b8b
93332ab7f7ff4141a371713871ab65ad 8568e0eab8364bfc89c876aadfa01022
43e14a86cbdd422880cac22d9a15d3c0 a0af8cb24ef94d25b9691eee1f7024ca 8ad...
INFO  - 2017-07-21 05:54:59.734; [contract

Few document replication not happen in solr cloud

2017-08-22 Thread Sanjay Lokhande


Hi guys,
I having 5 node solr cloud setup.  The logs indicates leader
and 2 solr node receiving document add request.  The other 2 nodes logs did
not show the entry to add the document and also these nodes missing the
document.
  - how such issue can be troubleshoot? any pointers?
 - any logging level which can help is this situation?

Regards,
Sanjay




Re: Few document replication not happen in solr cloud

2017-08-22 Thread Sanjay Lokhande
java:58)
at org.apache.solr.handler.UpdateRequestHandler$1.load
(UpdateRequestHandler.java:98)
at
org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody
(ContentStreamHandlerBase.java:74)
at org.apache.solr.handler.RequestHandlerBase.handleRequest
(RequestHandlerBase.java:143)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:2064)
at org.apache.solr.servlet.HttpSolrCall.execute
(HttpSolrCall.java:654)
at org.apache.solr.servlet.HttpSolrCall.call
(HttpSolrCall.java:450)
at org.apache.solr.servlet.SolrDispatchFilter.doFilter
(SolrDispatchFilter.java:227)
at org.apache.solr.servlet.SolrDispatchFilter.doFilter
(SolrDispatchFilter.java:196)
at org.eclipse.jetty.servlet.ServletHandler
$CachedChain.doFilter(ServletHandler.java:1652)
at org.eclipse.jetty.servlet.ServletHandler.doHandle
(ServletHandler.java:585)
at org.eclipse.jetty.server.handler.ScopedHandler.handle
(ScopedHandler.java:143)
at org.eclipse.jetty.security.SecurityHandler.handle
(SecurityHandler.java:577)
at org.eclipse.jetty.server.session.SessionHandler.doHandle
(SessionHandler.java:223)
at org.eclipse.jetty.server.handler.ContextHandler.doHandle
(ContextHandler.java:1127)
at org.eclipse.jetty.servlet.ServletHandler.doScope
(ServletHandler.java:515)
at org.eclipse.jetty.server.session.SessionHandler.doScope
(SessionHandler.java:185)
at org.eclipse.jetty.server.handler.ContextHandler.doScope
(ContextHandler.java:1061)
at org.eclipse.jetty.server.handler.ScopedHandler.handle
(ScopedHandler.java:141)
at
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle
(ContextHandlerCollection.java:215)
at org.eclipse.jetty.server.handler.HandlerCollection.handle
(HandlerCollection.java:110)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle
(HandlerWrapper.java:97)
at org.eclipse.jetty.server.Server.handle(Server.java:497)
at org.eclipse.jetty.server.HttpChannel.handle
(HttpChannel.java:310)
at org.eclipse.jetty.server.HttpConnection.onFillable
(HttpConnection.java:257)
at org.eclipse.jetty.io.AbstractConnection$2.run
(AbstractConnection.java:540)
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob
(QueuedThreadPool.java:635)
at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run
(QueuedThreadPool.java:555)
at java.lang.Thread.run(Thread.java:785)
Caused by: java.util.concurrent.TimeoutException: Idle timeout
expired: 5/5 ms
at org.eclipse.jetty.io.IdleTimeout.checkIdleTimeout
(IdleTimeout.java:161)
at org.eclipse.jetty.io.IdleTimeout$1.run(IdleTimeout.java:50)
at java.util.concurrent.Executors$RunnableAdapter.call
(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:267)
at java.util.concurrent.ScheduledThreadPoolExecutor
$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor
$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker
(ThreadPoolExecutor.java:1143)
at java.util.concurrent.ThreadPoolExecutor$Worker.run
(ThreadPoolExecutor.java:618)
... 1 more

Thanks in advance.

Regards,
Sanjay



From:   "Sanjay Lokhande" 
To: solr-user@lucene.apache.org
Date:   08/22/2017 09:00 PM
Subject:Few document replication not happen in solr cloud





Hi guys,
I having 5 node solr cloud setup.  The logs indicates leader
and 2 solr node receiving document add request.  The other 2 nodes logs did
not show the entry to add the document and also these nodes missing the
document.
  - how such issue can be troubleshoot? any pointers?
 - any logging level which can help is this situation?

Regards,
Sanjay






Re: Solr user interface

2008-07-10 Thread Sanjay N
Hi Lars,

Use solr with tomcat u will get interface from tomcat itself.

Regards
Sanjay

On Thu, Jul 10, 2008 at 4:32 PM, Lars Kotthoff <[EMAIL PROTECTED]> wrote:

> Hi all,
>
>  I've written a user interface for Solr (Spring web application) which I'd
> be
> willing to donate if people are interested.
>
> You can see a demo here http://larsko.dyndns.org:8080/solr-ui/search.html,
> SVN
> repository is here http://larsko.dyndns.org/svn/solr-ui/. Note in
> particular
> http://larsko.dyndns.org/svn/solr-ui/documentation/manual.pdf for a short
> manual. Please be patient, the server this is running on doesn't have a lot
> of
> processing power or upstream bandwidth ;)
>
> The purpose of adding this user interface to Solr would be twofold; first,
> serve
> as a demonstration of Solr's capabilities (running on a server linked to
> from
> the website, probably like the demo above), and second, give people a
> starting
> point/inspiration for implementing their own user interfaces.
>
> The special feature is that it supports some form of hierarchical faceting
> (explained in the manual). The data the demo searches comes from the
> wikipedia
> selection for schools. The subject index pages are used to build the
> hierarchy.
>
> Let me know what you think.
>
> Thanks,
>
> Lars
>


Re: java.io.IOException: read past EOF

2008-07-27 Thread Sanjay N
Hi Fuad,

We are using Server with 750 * 8 on Raid 5, all are segate Harddrives. & we
have enough space on Harddrive.

We are using Tomcat + solr

Regards
Sanjay



On Tue, Jul 22, 2008 at 9:19 PM, Fuad Efendi <[EMAIL PROTECTED]> wrote:

> Lucene index corrupted... which harddrive do you use?
>
>
> Quoting Rohan <[EMAIL PROTECTED]>:
>
>  Hi Guys,
>>
>> This is my first post. We are running solr with multiple Indexes, 20
>> Indexes. I'm facing problem with 5 one. I'm not able to run optimized on
>> that index. I'm getting following error. Your help is really appreciated.
>>
>>
>> java.io.IOException: read past EOF
>>at
>>
>> org.apache.lucene.store.BufferedIndexInput.refill(BufferedIndexInput.java:89
>>
>
>
>
>


Special character matching 'x' ?

2008-09-17 Thread Sanjay Suri
Hi,
Can someone shed some light on this?

One of my field values has  the name "Räikkönen"  which contains a special
characters.

Strangely, as I see it anyway, it matches on the search query 'x' ?

Can someone explain or point me to the solution/documentation?

Any help appreciated,
-Sanjay

-- 
Sanjay Suri

Videocrux Inc.
http://videocrux.com
+91 99102 66626


Re: Special character matching 'x' ?

2008-09-18 Thread Sanjay Suri
Thanks Akshay and Norberto,
I am still trying to make it work. I know the solution is what you pointed
me to but is just taking me some time to make it work.

thanks,
-Sanjay

On Thu, Sep 18, 2008 at 12:34 PM, Norberto Meijome <[EMAIL PROTECTED]>wrote:

> On Thu, 18 Sep 2008 10:53:39 +0530
> "Sanjay Suri" <[EMAIL PROTECTED]> wrote:
>
> > One of my field values has  the name "R__ikk__nen"  which contains a
> special
> > characters.
> >
> > Strangely, as I see it anyway, it matches on the search query 'x' ?
> >
> > Can someone explain or point me to the solution/documentation?
>
> hi Sanjay,
> Akshay should have given you an answer for this. In a more general way, if
> you
> want to know WHY something is matching the way it is, run the query with
> debugQuery=true . There are a few pages in the wiki which explain other
> debugging techniques.
>
> b
> _
> {Beto|Norberto|Numard} Meijome
>
> "Ask not what's inside your head, but what your head's inside of."
>   J. J. Gibson
>
> I speak for myself, not my employer. Contents may be hot. Slippery when
> wet.
> Reading disclaimers makes you go blind. Writing them is worse. You have
> been
> Warned.
>



-- 
Sanjay Suri

Videocrux Inc.
http://videocrux.com
+91 99102 66626


Newbie: For stopword query - All objects being returned

2008-11-20 Thread Sanjay Suri
Hi ,
I realize this might be too simple - Can someone tell me where to look? I'm
new to solr and have to fix this for a demo asap.

If my search query is "the", all 91 objects are returned as search results.
I expect 0 results.

-- 
Sanjay Suri

Videocrux Inc.
http://videocrux.com
+91 99102 66626


org.apache.lucene.util.fst.FST taking up lot of Java Heap Memory

2020-08-03 Thread sanjay dutt
Hello Solr community,
On our Production SolrCloud Server, OutOfMemory has been occurring on lot of 
instances. When I download the HEAP DUMP and analyzed it. I got to know that in 
multiple HEAP DUMPS there are lots of instances of 
org.apache.lucene.codecs.blocktree.BlockTreeTermsReader  which has the highest 
retained heap memory and further I have checked the outgoing-reference for 
those objects, the  org.apache.lucene.util.fst.FST is the one which occupy 90% 
of the heap memory.
it's like 
Production HEAP memory :- 12GBout of which  
org.apache.lucene.codecs.blocktree.BlockTreeTermsReader total retained heap :- 
7-8 GB(vary from instance to instance)and org.apache.lucene.util.fst.FST total 
retained heap :- 6-7 GB
Upon further looking I have calculated the total retained heap for 
FieldReader.fieldInfo.name="my_field" is around 7GB. Now this is the same 
reader which also contains reference to org.apache.lucene.util.fst.FST.
Now "my_field" is the field on which we are performing spatial searches. Is 
spatial searches use FST internally and hence we are seeing lot of heap memory 
used by FST.l only.
IS there any way we can optimize the spatial searches so that it take less 
memory.
Can someone please give me any pointer that from where Should I start looking 
to debug the above issue. 
Thanks and Regards,Sanjay Dutt
Sent from Yahoo Mail on Android

Sort on internal lucene ID led to OOM

2020-08-04 Thread sanjay dutt
Hello,
We were investigating one HEAP DUMP in which FielsCaceImpl has occupied around 
5GB. Upon further investigation, I got to know that 
FieldCacheImpl$SortedDocValues occupies 90% of the memory where 
FieldCacheImpl$CacheKey is "id".
When I checked the logs I was trying to find any query in which there is any 
sort on id field. But instead I find few queries in which we are sorting on 
_docid_ lucene internal id.
Can sort on _docid_ cause OOM?
And if I will enable docValues for uniqueKey(id) will that solve my problem?
Regards,SanjaySent from Yahoo Mail on Android

Re: Sort on internal lucene ID led to OOM

2020-08-05 Thread sanjay dutt
Thanks Erik.When you mentioned we can't control sorting by _docid_ . So if not 
client Does that mean solr internally makes such type of queries which includes 
sort on internal Lucene Id? And for what purpose?
Sent from Yahoo Mail on Android 
 
  On Wed, Aug 5, 2020 at 6:27 PM, Erick Erickson 
wrote:   A sort on anything can cause an OOM… That said, _all_ fields defined 
in your Solr schema should have docValues set to true if you sort, group, use 
function queries or facet on them. What’s happening is the docValues structure 
is being synthesized at runtime on the heap. In recent Solr releases you can 
specify uninvertible=true flag to generate exceptions when this is attempted 
rather than hit an OOM, see: 
https://lucene.apache.org/solr/guide/8_3/defining-fields.html

You can’t control sorting by _docid_, but if you’re specifying that from the 
client, it’s almost certainly not doing what you expect.

Best,
Erick

> On Aug 4, 2020, at 10:19 PM, sanjay dutt  
> wrote:
> 
> Hello,
> We were investigating one HEAP DUMP in which FielsCaceImpl has occupied 
> around 5GB. Upon further investigation, I got to know that 
> FieldCacheImpl$SortedDocValues occupies 90% of the memory where 
> FieldCacheImpl$CacheKey is "id".
> When I checked the logs I was trying to find any query in which there is any 
> sort on id field. But instead I find few queries in which we are sorting on 
> _docid_ lucene internal id.
> Can sort on _docid_ cause OOM?
> And if I will enable docValues for uniqueKey(id) will that solve my problem?
> Regards,SanjaySent from Yahoo Mail on Android
  


Re: org.apache.lucene.util.fst.FST taking up lot of Java Heap Memory

2020-08-06 Thread sanjay dutt
FieldType defined with class solr.SpatialRecursivePrefixTreeFieldType

In this we are adding points only although collection has few fields with 
points data and then other fieldTypes as well.
And one of the queries looks like 
(my_field: [45,-94 TO 46,-93]+OR+my_field: [42,-94 TO 43,-93])

Thanks and Regards,Sanjay Dutt 

On Thursday, August 6, 2020, 12:10:04 AM GMT+5:30, David Smiley 
 wrote:  
 
 What is the Solr field type definition for this field?  And what sort of
spatial data do you add here -- just points or what?

~ David Smiley
Apache Lucene/Solr Search Developer
http://www.linkedin.com/in/davidwsmiley


On Mon, Aug 3, 2020 at 10:09 PM sanjay dutt
 wrote:

> Hello Solr community,
> On our Production SolrCloud Server, OutOfMemory has been occurring on lot
> of instances. When I download the HEAP DUMP and analyzed it. I got to know
> that in multiple HEAP DUMPS there are lots of instances
> of org.apache.lucene.codecs.blocktree.BlockTreeTermsReader  which has the
> highest retained heap memory and further I have checked the
> outgoing-reference for those objects,
> the  org.apache.lucene.util.fst.FST is the one which occupy 90% of the heap
> memory.
> it's like
> Production HEAP memory :- 12GBout of
> which  org.apache.lucene.codecs.blocktree.BlockTreeTermsReader total retained
> heap :- 7-8 GB(vary from instance to
> instance)and org.apache.lucene.util.fst.FST total retained heap :- 6-7 GB
> Upon further looking I have calculated the total retained heap for
> FieldReader.fieldInfo.name="my_field" is around 7GB. Now this is the same
> reader which also contains reference to org.apache.lucene.util.fst.FST.
> Now "my_field" is the field on which we are performing spatial searches.
> Is spatial searches use FST internally and hence we are seeing lot of heap
> memory used by FST.l only.
> IS there any way we can optimize the spatial searches so that it take less
> memory.
> Can someone please give me any pointer that from where Should I start
> looking to debug the above issue.
> Thanks and Regards,Sanjay Dutt
> Sent from Yahoo Mail on Android
  

Re: org.apache.lucene.util.fst.FST taking up lot of Java Heap Memory

2020-08-07 Thread sanjay dutt
Best explanation I found so far. Will migrate to LatLonPointSpatialField and 
try to share the benchmark data here. Thanks again David.
Cheers,Sanjay

Sent from Yahoo Mail on Android 
 
  On Sat, Aug 8, 2020 at 3:31 AM, David Smiley wrote:   
Since you have a typical use-case (point data, queries that are
rectangles), I strongly encourage you to migrate to LatLonPointSpatialField:

https://builds.apache.org/job/Solr-reference-guide-master/javadoc/spatial-search.html#latlonpointspatialfield
It's based off an internal "BKD" tree index (doesn't use FSTs) which is
different than the terms based index used by the RPT field that you are
using which employes FSTs.  To be clear, FSTs are awesome but the BKD index
is tailored for numeric data whereas terms/FSTs are not.

If your FSTs are/were taking up so much memory, you are probably not using
Solr 8.4.0 or beyond, which moved to having the FSTs off-heap -- at least
the ones associated with the field indexes.

~ David Smiley
Apache Lucene/Solr Search Developer
http://www.linkedin.com/in/davidwsmiley


On Thu, Aug 6, 2020 at 8:19 PM sanjay dutt
 wrote:

> FieldType defined with class solr.SpatialRecursivePrefixTreeFieldType
>
> In this we are adding points only although collection has few fields with
> points data and then other fieldTypes as well.
> And one of the queries looks like
> (my_field: [45,-94 TO 46,-93]+OR+my_field: [42,-94 TO 43,-93])
>
> Thanks and Regards,Sanjay Dutt
>
>    On Thursday, August 6, 2020, 12:10:04 AM GMT+5:30, David Smiley <
> dsmi...@apache.org> wrote:
>
>  What is the Solr field type definition for this field?  And what sort of
> spatial data do you add here -- just points or what?
>
> ~ David Smiley
> Apache Lucene/Solr Search Developer
> http://www.linkedin.com/in/davidwsmiley
>
>
> On Mon, Aug 3, 2020 at 10:09 PM sanjay dutt
>  wrote:
>
> > Hello Solr community,
> > On our Production SolrCloud Server, OutOfMemory has been occurring on lot
> > of instances. When I download the HEAP DUMP and analyzed it. I got to
> know
> > that in multiple HEAP DUMPS there are lots of instances
> > of org.apache.lucene.codecs.blocktree.BlockTreeTermsReader  which has the
> > highest retained heap memory and further I have checked the
> > outgoing-reference for those objects,
> > the  org.apache.lucene.util.fst.FST is the one which occupy 90% of the
> heap
> > memory.
> > it's like
> > Production HEAP memory :- 12GBout of
> > which  org.apache.lucene.codecs.blocktree.BlockTreeTermsReader total
> retained
> > heap :- 7-8 GB(vary from instance to
> > instance)and org.apache.lucene.util.fst.FST total retained heap :- 6-7 GB
> > Upon further looking I have calculated the total retained heap for
> > FieldReader.fieldInfo.name="my_field" is around 7GB. Now this is the
> same
> > reader which also contains reference to org.apache.lucene.util.fst.FST.
> > Now "my_field" is the field on which we are performing spatial searches.
> > Is spatial searches use FST internally and hence we are seeing lot of
> heap
> > memory used by FST.l only.
> > IS there any way we can optimize the spatial searches so that it take
> less
> > memory.
> > Can someone please give me any pointer that from where Should I start
> > looking to debug the above issue.
> > Thanks and Regards,Sanjay Dutt
> > Sent from Yahoo Mail on Android
>
  


Re: Solr Down Issue

2020-08-09 Thread sanjay dutt
It could be OOM killer. Try to monitor it's heap and there is a script in bin 
which basically kills solr when OOM occurs.

Sent from Yahoo Mail on Android 
 
  On Sun, Aug 9, 2020 at 8:14 PM, Ben wrote:   Can you send 
solr logs?

Best,
Ben

On Sun, Aug 9, 2020, 9:55 AM Rashmi Jain  wrote:

> Hello Team,
>
>                I am Rashmi jain implemented solr on one of our site
> bookswagon.com. last 2-3 month we are facing
> strange issue, solr down suddenly without interrupting.  We check solr
> login and also check application logs but no clue found there regarding
> this.
>                We have implemented solr 7.4 on Java SE 10 and have index
> data of books around 28 million.
>                Also we are running solr on Windows server 2012 standard
> with 32 RAM.
>                Please help us on this.
>
> Regards,
> Rashmi
>
>
>
  


SPLITSHARD failed after running for hours

2020-08-12 Thread sanjay dutt
Hello Solr community,
We tried to split shard of one collection which contains 80M documents. After 
running for few hours it failed with the exception 
org.apache.solr.common.SolrException.
Upon further investigation, I found below exception
Caused by: java.util.concurrent.RejectedExecutionException: Task 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$1@4c995faa 
rejected from 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@4cea9dd8[Terminated,
 pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 
115769]: 

We just ended up with two more sub shards but with ZERO documents in them. Can 
someone please assist me here what exactly may have happened here. And if we 
have to try again what should we keep in our mind so that it will execute 
successfully.
Thanks,Sanjay


Sent from Yahoo Mail on Android

_version_ taking too much memory

2020-11-17 Thread sanjay dutt
Solr Collection setup
Shards :- 2Replication :- 4Documents :- 569 Million (I know it's too much)Heap 
Space :- 12GB
So basically, above collection is having OutOfMemory issues. And upon 
inspection, I got to know that 
org.apache.lucene.uninverting.FieldCacheImpl$LongsFromArray for field 
"_version_" is taking around 50 - 60% memory(5GB).
Now I am trying to understand How "_version_" internally using the field cache?
Around the same time, In Splunk logs I can see "Found MAX value from Terms for 
_version_ in index". and also few "firstSearcher" queries. Now I think upon 
core initialization "firstSearcher" triggered which then try to find the Max 
value for version using Field Cache but because of too many records it occupy 
lot of memory to carry out the task.
Can someone please validate Is what I am talking in above para is the problem 
or problem lies in somewhere else?

Best Regards,Sanjay


Re: _version_ taking too much memory

2020-11-18 Thread sanjay dutt
Thank you so much Shalin. Yes that would reduce the memory requirement at run 
time.But my next question, Is the "firstSearcher" doing this I mean due to 
which it tries to find the "Max value of version" and for that it it has to do 
the un-inversion. 
Thanks and Regards,Sanjay Dutt 

On Wednesday, November 18, 2020, 12:54:39 PM GMT+5:30, Shalin Shekhar 
Mangar  wrote:  
 
 You should change the _version_ field to be of type docValues and re-index
data. This way you don't have to pay the memory cost of un-inversion.

On Wed, Nov 18, 2020 at 9:51 AM sanjay dutt
 wrote:

> Solr Collection setup
> Shards :- 2Replication :- 4Documents :- 569 Million (I know it's too
> much)Heap Space :- 12GB
> So basically, above collection is having OutOfMemory issues. And upon
> inspection, I got to know
> that org.apache.lucene.uninverting.FieldCacheImpl$LongsFromArray for field
> "_version_" is taking around 50 - 60% memory(5GB).
> Now I am trying to understand How "_version_" internally using the field
> cache?
> Around the same time, In Splunk logs I can see "Found MAX value from Terms
> for _version_ in index". and also few "firstSearcher" queries. Now I think
> upon core initialization "firstSearcher" triggered which then try to find
> the Max value for version using Field Cache but because of too many records
> it occupy lot of memory to carry out the task.
> Can someone please validate Is what I am talking in above para is the
> problem or problem lies in somewhere else?
>
> Best Regards,Sanjay
>


-- 
Regards,
Shalin Shekhar Mangar.
  

Help with Solr File Based spell check

2012-06-20 Thread Sanjay Dua - Network
Hi,

We are trying to implement file based search in our application using Solr 1.4. 
This is the code we have written

-<http://lllydevvm02.sixfeetup.com:8983/solr/admin/file/?file=solrconfig.xml> 

-<http://lllydevvm02.sixfeetup.com:8983/solr/admin/file/?file=solrconfig.xml> 

  default
  solr.FileBasedSpellChecker
  /usr/home/lilly/sixfeetup/projects/alm-buildout/etc/solr/spelling.txt
  ./filespellchecker
  0.7
  
  text
  


We are facing a issue and need your help on the same.

When the user searches for a word "medicine", which is a correct word and is 
present in the dictionary. We still get a suggestion "medicines" from 
dictionary.

We only want suggestion if the word is incorrectly spelled or is not included 
in the dictionary.

Can you please provide some suggestions.

Regards,
Sanjay Dua