Solr Relevancy problem

2020-02-19 Thread Pradeep Tambade
Hello,

We have configured solr site search engine into our website(www.croma.com). We 
are facing various issues like not showing relevant results, free text search 
not showing  result, phrase keywords shows irrelevant results etc

Please help us resolve these issues also help to connect with solr tech support 
team or any other company who is expert in managing solr search.


Thanks & Regards,
Pradeep Tambade |  Assistant Manager - Business Analyst
Infiniti Retail Ltd. | A Tata Enterprise
Mobile: +91 9664536737
Email: pradeep.tamb...@croma.com | Shop at: www.croma.com


  Have e-waste but don't know what to do about it?

  *   Call us at 7207-666-000 & we pick up your junk at your doorstep
  *   We ensure responsible disposal
  *   And also plant an actual tree in your name for the e-waste you dispose

 [https://www.croma.com/_ui/responsive/common/images/Greatplacetowork_.jpg]

Registered Office: Unit No. 701 & 702, 7th Floor, Kaledonia, Sahar Road, 
Andheri East, Mumbai - 400069, India


Re: Solr Relevancy problem

2020-02-19 Thread Jörn Franke
The best way to address this problem is to collect queries and examples why 
they are wrong and to document this. This is especially important when working 
with another vendor. Otherwise no one can give you proper help.

> Am 19.02.2020 um 09:17 schrieb Pradeep Tambade 
> :
> 
> Hello,
> 
> We have configured solr site search engine into our website(www.croma.com). 
> We are facing various issues like not showing relevant results, free text 
> search not showing  result, phrase keywords shows irrelevant results etc
> 
> Please help us resolve these issues also help to connect with solr tech 
> support team or any other company who is expert in managing solr search.
> 
> 
> Thanks & Regards,
> Pradeep Tambade |  Assistant Manager - Business Analyst
> Infiniti Retail Ltd. | A Tata Enterprise
> Mobile: +91 9664536737
> Email: pradeep.tamb...@croma.com | Shop at: www.croma.com
> 
> 
>  Have e-waste but don't know what to do about it?
> 
>  *   Call us at 7207-666-000 & we pick up your junk at your doorstep
>  *   We ensure responsible disposal
>  *   And also plant an actual tree in your name for the e-waste you dispose
> 
> [https://www.croma.com/_ui/responsive/common/images/Greatplacetowork_.jpg]
> 
> Registered Office: Unit No. 701 & 702, 7th Floor, Kaledonia, Sahar Road, 
> Andheri East, Mumbai - 400069, India


Re: Solr Relevancy problem

2020-02-19 Thread Atita Arora
+1 for John's reply.
Along with that you can try the debug query to see how the query is being
parsed and what's going wrong.

Hope it helps,
Atita

On Wed, 19 Feb 2020, 09:19 Jörn Franke,  wrote:

> The best way to address this problem is to collect queries and examples
> why they are wrong and to document this. This is especially important when
> working with another vendor. Otherwise no one can give you proper help.
>
> > Am 19.02.2020 um 09:17 schrieb Pradeep Tambade <
> pradeep.tamb...@croma.com.invalid>:
> >
> > Hello,
> >
> > We have configured solr site search engine into our website(
> www.croma.com). We are facing various issues like not showing relevant
> results, free text search not showing  result, phrase keywords shows
> irrelevant results etc
> >
> > Please help us resolve these issues also help to connect with solr tech
> support team or any other company who is expert in managing solr search.
> >
> >
> > Thanks & Regards,
> > Pradeep Tambade |  Assistant Manager - Business Analyst
> > Infiniti Retail Ltd. | A Tata Enterprise
> > Mobile: +91 9664536737
> > Email: pradeep.tamb...@croma.com | Shop at: www.croma.com
> >
> >
> >  Have e-waste but don't know what to do about it?
> >
> >  *   Call us at 7207-666-000 & we pick up your junk at your doorstep
> >  *   We ensure responsible disposal
> >  *   And also plant an actual tree in your name for the e-waste you
> dispose
> >
> > [
> https://www.croma.com/_ui/responsive/common/images/Greatplacetowork_.jpg]
> >
> > Registered Office: Unit No. 701 & 702, 7th Floor, Kaledonia, Sahar Road,
> Andheri East, Mumbai - 400069, India
>


RE: Solr Upgrade socketTimeout issue in 8.2

2020-02-19 Thread Yogesh Chaudhari
Hi,

Could you please share me the steps to upgrade SOlr?

Now I am using Solr cloud 5.2.1 on production and wanted to upgrade to 
SOlr7.7.2. I am doing this in 2 spteps SOlr 5.2.1 to SOlr 6.6.6 then SOlr 7.7.2.

I have upgraded to Solr but getting issue for indexing of old documents.  I am 
badly stuck get get old document in migrated solr version.

Should I do the re-indexing? If yes can you please share the way to re-indexing?

Can you please provide your inputs on this? 

Thanks,

Yogesh Chaudhari

-Original Message-
From: kshitij tyagi  
Sent: Wednesday, February 19, 2020 12:52 PM
To: solr-user@lucene.apache.org
Subject: Solr Upgrade socketTimeout issue in 8.2

Hi,

We have upgraded our solrCloud from version 6.6.0 to 8.2.0

At the time of indexing intermittently we are observing socketTimeout exception 
when using Collection apis. example when we try reloading one of the collection 
using CloudSolrClient class.

Is there any performance degradation in Solrcloud collection apis?

logs:

IOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@357] - caught end of 
stream exception

EndOfStreamException: Unable to read additional data from client sessionid 
0x2663e756d775747, likely client has closed socket

at org.apache.zookeeper.server.NIOServerCnxn.doIO(NIOServerCnxn.java:228)

at
org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:203)

at java.lang.Thread.run(Unknown Source)


logs:


Exception has occured in job switch: Timeout occurred while waiting response 
from server 
at:http://secure-web.cisco.com/1w_RA10DqbruLQVC6aKUXuMgV4hC3T14viv2m2iUptQ2hyGjYLn0sSSy0Q_XNqcVxHym-e_mOyPc_AYM4zpIWWXdyRCpvXzL3mSeFK-DzhL_CqoNi2FwQUvhk2zb8OQKs1e11yBHqblc3Kyx0XlruLvb24BUj0lBBGmJVf5E9rrTaFQbFmCdNyccx1KCIpzf2MlyeqvvXVWKCW_YbqnLWGjcfqlAylbNqJTGuKf5rbBMdJ8pn14dbFlM0QDZjn6IORWVA8NqmdhC9VwD1rzpU6dVIpsph6qz_OcgoH61wlZALQ1Zj65XRFtXvuhqEWQeaabvKactprjz1o3pflKaxttbgxz1ItRxb4FjZkBgTC24uwalAmi_CyfeP7DECtIYATYf3AJFjCUfLV8_Rj2V5J0JeCTFDi7CWqKFUhiHXtpM8PvZt8kgMIRwfgPUKHPIJ/http%3A%2F%2Fprod-t-8.net%3A8983%2Fsolr


Is anyone facing same type of issue? in Solrcloud? Any suggestions to solve??



Regards,

kshitij


Re: Solr Upgrade socketTimeout issue in 8.2

2020-02-19 Thread Jörn Franke
Yes you need to reindex.
Update solrconfig, schemas to leverage the later feature of the version (some 
datatypes are now more optimal others are deprecated.

Update Solrconfig.xml and schema to leverage the latest  datatypes , features 
etc..

Create new collection based on newest config.
Use your regular Index process to move documents to new collection.

Check if new collection works and has expected performance.

Delete old collection.

Test before in a test environment and not in production!

> Am 19.02.2020 um 09:46 schrieb Yogesh Chaudhari 
> :
> 
> Hi,
> 
> Could you please share me the steps to upgrade SOlr?
> 
> Now I am using Solr cloud 5.2.1 on production and wanted to upgrade to 
> SOlr7.7.2. I am doing this in 2 spteps SOlr 5.2.1 to SOlr 6.6.6 then SOlr 
> 7.7.2.
> 
> I have upgraded to Solr but getting issue for indexing of old documents.  I 
> am badly stuck get get old document in migrated solr version.
> 
> Should I do the re-indexing? If yes can you please share the way to 
> re-indexing?
> 
> Can you please provide your inputs on this? 
> 
> Thanks,
> 
> Yogesh Chaudhari
> 
> -Original Message-
> From: kshitij tyagi  
> Sent: Wednesday, February 19, 2020 12:52 PM
> To: solr-user@lucene.apache.org
> Subject: Solr Upgrade socketTimeout issue in 8.2
> 
> Hi,
> 
> We have upgraded our solrCloud from version 6.6.0 to 8.2.0
> 
> At the time of indexing intermittently we are observing socketTimeout 
> exception when using Collection apis. example when we try reloading one of 
> the collection using CloudSolrClient class.
> 
> Is there any performance degradation in Solrcloud collection apis?
> 
> logs:
> 
> IOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@357] - caught end of 
> stream exception
> 
> EndOfStreamException: Unable to read additional data from client sessionid 
> 0x2663e756d775747, likely client has closed socket
> 
> at org.apache.zookeeper.server.NIOServerCnxn.doIO(NIOServerCnxn.java:228)
> 
> at
> org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:203)
> 
> at java.lang.Thread.run(Unknown Source)
> 
> 
> logs:
> 
> 
> Exception has occured in job switch: Timeout occurred while waiting response 
> from server 
> at:http://secure-web.cisco.com/1w_RA10DqbruLQVC6aKUXuMgV4hC3T14viv2m2iUptQ2hyGjYLn0sSSy0Q_XNqcVxHym-e_mOyPc_AYM4zpIWWXdyRCpvXzL3mSeFK-DzhL_CqoNi2FwQUvhk2zb8OQKs1e11yBHqblc3Kyx0XlruLvb24BUj0lBBGmJVf5E9rrTaFQbFmCdNyccx1KCIpzf2MlyeqvvXVWKCW_YbqnLWGjcfqlAylbNqJTGuKf5rbBMdJ8pn14dbFlM0QDZjn6IORWVA8NqmdhC9VwD1rzpU6dVIpsph6qz_OcgoH61wlZALQ1Zj65XRFtXvuhqEWQeaabvKactprjz1o3pflKaxttbgxz1ItRxb4FjZkBgTC24uwalAmi_CyfeP7DECtIYATYf3AJFjCUfLV8_Rj2V5J0JeCTFDi7CWqKFUhiHXtpM8PvZt8kgMIRwfgPUKHPIJ/http%3A%2F%2Fprod-t-8.net%3A8983%2Fsolr
> 
> 
> Is anyone facing same type of issue? in Solrcloud? Any suggestions to solve??
> 
> 
> 
> Regards,
> 
> kshitij


Re: Solr Upgrade socketTimeout issue in 8.2

2020-02-19 Thread kshitij tyagi
Hi,

Any information on socket timeout issue when using collection apis? I am
observing increased response time when using Collection APis in upgraded
version

On Wed, Feb 19, 2020 at 2:22 PM Jörn Franke  wrote:

> Yes you need to reindex.
> Update solrconfig, schemas to leverage the later feature of the version
> (some datatypes are now more optimal others are deprecated.
>
> Update Solrconfig.xml and schema to leverage the latest  datatypes ,
> features etc..
>
> Create new collection based on newest config.
> Use your regular Index process to move documents to new collection.
>
> Check if new collection works and has expected performance.
>
> Delete old collection.
>
> Test before in a test environment and not in production!
>
> > Am 19.02.2020 um 09:46 schrieb Yogesh Chaudhari
> :
> >
> > Hi,
> >
> > Could you please share me the steps to upgrade SOlr?
> >
> > Now I am using Solr cloud 5.2.1 on production and wanted to upgrade to
> SOlr7.7.2. I am doing this in 2 spteps SOlr 5.2.1 to SOlr 6.6.6 then SOlr
> 7.7.2.
> >
> > I have upgraded to Solr but getting issue for indexing of old
> documents.  I am badly stuck get get old document in migrated solr version.
> >
> > Should I do the re-indexing? If yes can you please share the way to
> re-indexing?
> >
> > Can you please provide your inputs on this?
> >
> > Thanks,
> >
> > Yogesh Chaudhari
> >
> > -Original Message-
> > From: kshitij tyagi 
> > Sent: Wednesday, February 19, 2020 12:52 PM
> > To: solr-user@lucene.apache.org
> > Subject: Solr Upgrade socketTimeout issue in 8.2
> >
> > Hi,
> >
> > We have upgraded our solrCloud from version 6.6.0 to 8.2.0
> >
> > At the time of indexing intermittently we are observing socketTimeout
> exception when using Collection apis. example when we try reloading one of
> the collection using CloudSolrClient class.
> >
> > Is there any performance degradation in Solrcloud collection apis?
> >
> > logs:
> >
> > IOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@357] - caught
> end of stream exception
> >
> > EndOfStreamException: Unable to read additional data from client
> sessionid 0x2663e756d775747, likely client has closed socket
> >
> > at org.apache.zookeeper.server.NIOServerCnxn.doIO(NIOServerCnxn.java:228)
> >
> > at
> >
> org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:203)
> >
> > at java.lang.Thread.run(Unknown Source)
> >
> >
> > logs:
> >
> >
> > Exception has occured in job switch: Timeout occurred while waiting
> response from server at:
> http://secure-web.cisco.com/1w_RA10DqbruLQVC6aKUXuMgV4hC3T14viv2m2iUptQ2hyGjYLn0sSSy0Q_XNqcVxHym-e_mOyPc_AYM4zpIWWXdyRCpvXzL3mSeFK-DzhL_CqoNi2FwQUvhk2zb8OQKs1e11yBHqblc3Kyx0XlruLvb24BUj0lBBGmJVf5E9rrTaFQbFmCdNyccx1KCIpzf2MlyeqvvXVWKCW_YbqnLWGjcfqlAylbNqJTGuKf5rbBMdJ8pn14dbFlM0QDZjn6IORWVA8NqmdhC9VwD1rzpU6dVIpsph6qz_OcgoH61wlZALQ1Zj65XRFtXvuhqEWQeaabvKactprjz1o3pflKaxttbgxz1ItRxb4FjZkBgTC24uwalAmi_CyfeP7DECtIYATYf3AJFjCUfLV8_Rj2V5J0JeCTFDi7CWqKFUhiHXtpM8PvZt8kgMIRwfgPUKHPIJ/http%3A%2F%2Fprod-t-8.net%3A8983%2Fsolr
> >
> >
> > Is anyone facing same type of issue? in Solrcloud? Any suggestions to
> solve??
> >
> >
> >
> > Regards,
> >
> > kshitij
>


Re: Solr Relevancy problem

2020-02-19 Thread Paras Lehana
Hi Pradeep,

I suggest you to at least post an example query, its result and the result
you were expecting. How do you boost your documents?

I guess croma has just started using Solr (I could not find you here
) and if
that's the case, don't worry - Solr is very powerful and it takes time
initially to arrive at better queries.

On Wed, 19 Feb 2020 at 13:56, Atita Arora  wrote:

> +1 for John's reply.
> Along with that you can try the debug query to see how the query is being
> parsed and what's going wrong.
>
> Hope it helps,
> Atita
>
> On Wed, 19 Feb 2020, 09:19 Jörn Franke,  wrote:
>
> > The best way to address this problem is to collect queries and examples
> > why they are wrong and to document this. This is especially important
> when
> > working with another vendor. Otherwise no one can give you proper help.
> >
> > > Am 19.02.2020 um 09:17 schrieb Pradeep Tambade <
> > pradeep.tamb...@croma.com.invalid>:
> > >
> > > Hello,
> > >
> > > We have configured solr site search engine into our website(
> > www.croma.com). We are facing various issues like not showing relevant
> > results, free text search not showing  result, phrase keywords shows
> > irrelevant results etc
> > >
> > > Please help us resolve these issues also help to connect with solr tech
> > support team or any other company who is expert in managing solr search.
> > >
> > >
> > > Thanks & Regards,
> > > Pradeep Tambade |  Assistant Manager - Business Analyst
> > > Infiniti Retail Ltd. | A Tata Enterprise
> > > Mobile: +91 9664536737
> > > Email: pradeep.tamb...@croma.com | Shop at: www.croma.com
> > >
> > >
> > >  Have e-waste but don't know what to do about it?
> > >
> > >  *   Call us at 7207-666-000 & we pick up your junk at your doorstep
> > >  *   We ensure responsible disposal
> > >  *   And also plant an actual tree in your name for the e-waste you
> > dispose
> > >
> > > [
> > https://www.croma.com/_ui/responsive/common/images/Greatplacetowork_.jpg
> ]
> > >
> > > Registered Office: Unit No. 701 & 702, 7th Floor, Kaledonia, Sahar
> Road,
> > Andheri East, Mumbai - 400069, India
> >
>


-- 
-- 
Regards,

*Paras Lehana* [65871]
Development Engineer, *Auto-Suggest*,
IndiaMART InterMESH Ltd,

11th Floor, Tower 2, Assotech Business Cresterra,
Plot No. 22, Sector 135, Noida, Uttar Pradesh, India 201305

Mob.: +91-9560911996
Work: 0120-4056700 | Extn:
*11096*

-- 
*
*

 


Re: Solr Relevancy problem

2020-02-19 Thread Erick Erickson
Please review:

https://cwiki.apache.org/confluence/display/solr/UsingMailingLists

> On Feb 19, 2020, at 7:26 AM, Paras Lehana  wrote:
> 
> Hi Pradeep,
> 
> I suggest you to at least post an example query, its result and the result
> you were expecting. How do you boost your documents?
> 
> I guess croma has just started using Solr (I could not find you here
> ) and if
> that's the case, don't worry - Solr is very powerful and it takes time
> initially to arrive at better queries.
> 
> On Wed, 19 Feb 2020 at 13:56, Atita Arora  wrote:
> 
>> +1 for John's reply.
>> Along with that you can try the debug query to see how the query is being
>> parsed and what's going wrong.
>> 
>> Hope it helps,
>> Atita
>> 
>> On Wed, 19 Feb 2020, 09:19 Jörn Franke,  wrote:
>> 
>>> The best way to address this problem is to collect queries and examples
>>> why they are wrong and to document this. This is especially important
>> when
>>> working with another vendor. Otherwise no one can give you proper help.
>>> 
 Am 19.02.2020 um 09:17 schrieb Pradeep Tambade <
>>> pradeep.tamb...@croma.com.invalid>:
 
 Hello,
 
 We have configured solr site search engine into our website(
>>> www.croma.com). We are facing various issues like not showing relevant
>>> results, free text search not showing  result, phrase keywords shows
>>> irrelevant results etc
 
 Please help us resolve these issues also help to connect with solr tech
>>> support team or any other company who is expert in managing solr search.
 
 
 Thanks & Regards,
 Pradeep Tambade |  Assistant Manager - Business Analyst
 Infiniti Retail Ltd. | A Tata Enterprise
 Mobile: +91 9664536737
 Email: pradeep.tamb...@croma.com | Shop at: www.croma.com
 
 
 Have e-waste but don't know what to do about it?
 
 *   Call us at 7207-666-000 & we pick up your junk at your doorstep
 *   We ensure responsible disposal
 *   And also plant an actual tree in your name for the e-waste you
>>> dispose
 
 [
>>> https://www.croma.com/_ui/responsive/common/images/Greatplacetowork_.jpg
>> ]
 
 Registered Office: Unit No. 701 & 702, 7th Floor, Kaledonia, Sahar
>> Road,
>>> Andheri East, Mumbai - 400069, India
>>> 
>> 
> 
> 
> -- 
> -- 
> Regards,
> 
> *Paras Lehana* [65871]
> Development Engineer, *Auto-Suggest*,
> IndiaMART InterMESH Ltd,
> 
> 11th Floor, Tower 2, Assotech Business Cresterra,
> Plot No. 22, Sector 135, Noida, Uttar Pradesh, India 201305
> 
> Mob.: +91-9560911996
> Work: 0120-4056700 | Extn:
> *11096*
> 
> -- 
> *
> *
> 
> 



Re: Solr Upgrade socketTimeout issue in 8.2

2020-02-19 Thread Erick Erickson
Yogesh:

Please do not hijack threads. The original poster requested information about
socket timeouts. True “upgrade” was mentioned, but it was a completely
different issue.

Kshitij:

There’s not much information to go on here. It’s possible you were running
close to the timeout limit before and “something” changed just enough
to go over that limit.

I’m a bit confused though, you talk about commands like reload while indexing.
What _exactly_ are you trying to do? Details matter.

One thing that did changes was “schemaless” became the default. This
causes reloads when Solr is indexing docs and comes across fields
for the first time. I personally don’t recommend “schemaless”, so what
happens if you turn that off?

If you’re manually sending reloads, you might try doing them async.

That said, you must start from a fresh index with _no_ documents in it
when you upgrade more than one major version. Did you start over?

Best,
Erick

> On Feb 19, 2020, at 3:58 AM, kshitij tyagi  wrote:
> 
> Hi,
> 
> Any information on socket timeout issue when using collection apis? I am
> observing increased response time when using Collection APis in upgraded
> version
> 
> On Wed, Feb 19, 2020 at 2:22 PM Jörn Franke  wrote:
> 
>> Yes you need to reindex.
>> Update solrconfig, schemas to leverage the later feature of the version
>> (some datatypes are now more optimal others are deprecated.
>> 
>> Update Solrconfig.xml and schema to leverage the latest  datatypes ,
>> features etc..
>> 
>> Create new collection based on newest config.
>> Use your regular Index process to move documents to new collection.
>> 
>> Check if new collection works and has expected performance.
>> 
>> Delete old collection.
>> 
>> Test before in a test environment and not in production!
>> 
>>> Am 19.02.2020 um 09:46 schrieb Yogesh Chaudhari
>> :
>>> 
>>> Hi,
>>> 
>>> Could you please share me the steps to upgrade SOlr?
>>> 
>>> Now I am using Solr cloud 5.2.1 on production and wanted to upgrade to
>> SOlr7.7.2. I am doing this in 2 spteps SOlr 5.2.1 to SOlr 6.6.6 then SOlr
>> 7.7.2.
>>> 
>>> I have upgraded to Solr but getting issue for indexing of old
>> documents.  I am badly stuck get get old document in migrated solr version.
>>> 
>>> Should I do the re-indexing? If yes can you please share the way to
>> re-indexing?
>>> 
>>> Can you please provide your inputs on this?
>>> 
>>> Thanks,
>>> 
>>> Yogesh Chaudhari
>>> 
>>> -Original Message-
>>> From: kshitij tyagi 
>>> Sent: Wednesday, February 19, 2020 12:52 PM
>>> To: solr-user@lucene.apache.org
>>> Subject: Solr Upgrade socketTimeout issue in 8.2
>>> 
>>> Hi,
>>> 
>>> We have upgraded our solrCloud from version 6.6.0 to 8.2.0
>>> 
>>> At the time of indexing intermittently we are observing socketTimeout
>> exception when using Collection apis. example when we try reloading one of
>> the collection using CloudSolrClient class.
>>> 
>>> Is there any performance degradation in Solrcloud collection apis?
>>> 
>>> logs:
>>> 
>>> IOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@357] - caught
>> end of stream exception
>>> 
>>> EndOfStreamException: Unable to read additional data from client
>> sessionid 0x2663e756d775747, likely client has closed socket
>>> 
>>> at org.apache.zookeeper.server.NIOServerCnxn.doIO(NIOServerCnxn.java:228)
>>> 
>>> at
>>> 
>> org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:203)
>>> 
>>> at java.lang.Thread.run(Unknown Source)
>>> 
>>> 
>>> logs:
>>> 
>>> 
>>> Exception has occured in job switch: Timeout occurred while waiting
>> response from server at:
>> http://secure-web.cisco.com/1w_RA10DqbruLQVC6aKUXuMgV4hC3T14viv2m2iUptQ2hyGjYLn0sSSy0Q_XNqcVxHym-e_mOyPc_AYM4zpIWWXdyRCpvXzL3mSeFK-DzhL_CqoNi2FwQUvhk2zb8OQKs1e11yBHqblc3Kyx0XlruLvb24BUj0lBBGmJVf5E9rrTaFQbFmCdNyccx1KCIpzf2MlyeqvvXVWKCW_YbqnLWGjcfqlAylbNqJTGuKf5rbBMdJ8pn14dbFlM0QDZjn6IORWVA8NqmdhC9VwD1rzpU6dVIpsph6qz_OcgoH61wlZALQ1Zj65XRFtXvuhqEWQeaabvKactprjz1o3pflKaxttbgxz1ItRxb4FjZkBgTC24uwalAmi_CyfeP7DECtIYATYf3AJFjCUfLV8_Rj2V5J0JeCTFDi7CWqKFUhiHXtpM8PvZt8kgMIRwfgPUKHPIJ/http%3A%2F%2Fprod-t-8.net%3A8983%2Fsolr
>>> 
>>> 
>>> Is anyone facing same type of issue? in Solrcloud? Any suggestions to
>> solve??
>>> 
>>> 
>>> 
>>> Regards,
>>> 
>>> kshitij
>> 



Re: Solr Relevancy problem

2020-02-19 Thread Jan Høydahl
Sounds like you really want an expert to work with you, in which case you can 
check https://cwiki.apache.org/confluence/display/SOLR/Support for a list of 
companies that claims to known Solr :)

Jan

> 19. feb. 2020 kl. 05:22 skrev Pradeep Tambade 
> :
> 
> Hello,
> 
> We have configured solr site search engine into our website(www.croma.com). 
> We are facing various issues like not showing relevant results, free text 
> search not showing  result, phrase keywords shows irrelevant results etc
> 
> Please help us resolve these issues also help to connect with solr tech 
> support team or any other company who is expert in managing solr search.
> 
> 
> Thanks & Regards,
> Pradeep Tambade |  Assistant Manager - Business Analyst
> Infiniti Retail Ltd. | A Tata Enterprise
> Mobile: +91 9664536737
> Email: pradeep.tamb...@croma.com | Shop at: www.croma.com
> 
> 
>  Have e-waste but don't know what to do about it?
> 
>  *   Call us at 7207-666-000 & we pick up your junk at your doorstep
>  *   We ensure responsible disposal
>  *   And also plant an actual tree in your name for the e-waste you dispose
> 
> [https://www.croma.com/_ui/responsive/common/images/Greatplacetowork_.jpg]
> 
> Registered Office: Unit No. 701 & 702, 7th Floor, Kaledonia, Sahar Road, 
> Andheri East, Mumbai - 400069, India



Re: Bug? Documents not visible after sucessful commit - chaos testing

2020-02-19 Thread Michael Frank
Hi Chriss,
thanks for opening the ticket. I have found some possibly related issues:
Open:
https://issues.apache.org/jira/browse/SOLR-3888  - "need beter handling of
external add/commit requests during tlog recovery"


Closed:
https://issues.apache.org/jira/browse/SOLR-12011
https://issues.apache.org/jira/browse/SOLR-9366

Cheers,
Michael

Am Do., 13. Feb. 2020 um 19:19 Uhr schrieb Chris Hostetter <
hossman_luc...@fucit.org>:

>
> : We think this is a bug (silently dropping commits even if the client
> : requested "waitForSearcher"), or at least a missing feature (commits
> beging
> : the only UpdateRequests not reporting the achieved RF), which should be
> : worth a JIRA Ticket.
>
> Thanks for your analysis Michael -- I agree something better should be
> done here, and have filed SOLR-14262 for subsequent discussion...
>
> https://issues.apache.org/jira/browse/SOLR-14262
>
> I believe the reason the local commit is ignored during replay is to
> ensure a consistent view of the index -- if the tlog being
> replayed contains COMMIT1,A,B,C,COMMIT2,D,... we should never open a new
> searcher containing just A or just A+B w/o C if a COMMIT3 comes along
> during replay -- but agree with you 100% that either commit should support
> 'rf' making it obvious that this commit didn't succeed (which would also
> be important & helpful if the node was still down when the client sends
> the commit) ... *AND* ... we should consider making the commit block until
> replay is finished.
>
> ...BUT... there are probably other nuances i don't understand ...
> hoepfully other folks more familiar with the current implementation will
> chime in on the jira.
>
>
>
>
> -Hoss
> http://www.lucidworks.com/
>


Phrase search and WordDelimiterGraphFilter not working as expected with mixed delimited and non-delimited tokens

2020-02-19 Thread Jeroen Steggink | knowsy

Hi,

I have a question regarding phrase search in combination with a 
WordDelimiterGraphFilter (Solr 8.4.1).


Whenever I try to search using a phrase where token combination consists 
of delimited and non-delimited tokens, I don't get any matches.


This is the configuration:


  





  
  



  





Example document:

{
  id: '1',
  text: 'mr. i.n.i.t. firstsirname secondsirname'
}

Queries and results:

Query:
"mr. i.n.i.t. firstsirname"
-
No result

Query:
"mr. i.n.i.t."
-
Result

Query:
"mr. i n i t"
-
Result

Query:
"mr. init"
-
Result

Query:
"mr init"
-
Result

Query:
"i.n.i.t. firstsirname"
-
No result

Query:
"init firstsirname"
-
No result

Query:
"i.n.i.t. firstsirname secondsirname"
-
No result

Query:
"init firstsirname secondsirname"
-
No result


I don't quite understand why this is. When looking at the results of the 
analyzers I don't understand why it's working with just delimited or 
non-delimited tokens. However, as soon as the mixed combination of 
delimited and non-delimited is searched, there is no match.


Could someone explain? And is there a solution to make it work?

Best regards,

Jeroen




How to auto-scale Solr and maintain even distribution of shard replicas across the whole cluster?

2020-02-19 Thread Christian Platta
Hi,
I'm playing around with the autoscaling feature of Solr 7.7.1 and have the 
following scenario to solve:

- One collection with two shards
- I want to activate autoscaling to achieve the following behavior:
  * Every time a new node comes up, it should get a new replica automatically 
through the autoscaling feature of Solr
  * Each node should contain only 1 replica (either from shard1 or shard2)
  * Solr should try to balance the number of replica in the cluster if 
possible, meaning:
 * if I have an even number of nodes in the cluster half of the nodes 
should have a replica for shard1 and the other half should have a replica for 
shard2
 * if I have an odd number of nodes one of the two shards has one replica 
more than the other shard

The scaling should look like this:
1) node a comes up and gets a shard1 replica
2) node b comes up and gets a shard2 replica
3) node c comes up and gets a shard1 replica
4) node d comes up and gets a shard2 replica
5) etc.

The problem I have is that if a new node comes up it always gets a replica from 
the second shard.
There is no balancing between the number of replicas of the two shards.

I have created the following autoscaling policies:

curl -X POST -H 'Content-Type: application/json' \
"${SOLR_LOCAL_URL}/api/cluster/autoscaling" --data-binary '{
"set-cluster-policy": [
{"replica": "<2", "shard": "#ANY", "node": "#ANY"},
{"replica": "#EQUAL", "shard": "#ANY", "node": "#ANY"}
]
}'

It seems that all policies are always evaluated from the point of view of a 
single node and not from the point of view of the whole cluster (it seems as if 
node:#ANY means every single node on its own and not the whole cluster).

Can my desired scaling behavior be achieved with the current autoscaling 
implementation in Solr?

My cluster setup looks like this:

#!/bin/bash
SOLR_LOCAL_URL=http://localhost:9000

# node added trigger
curl --header "Content-Type: application/json" \
   --request POST \
   --data '{ "set-trigger": { "name": "node_added_trigger", "event": 
"nodeAdded", "waitFor": "10s", "enabled": "true", "preferredOperation": 
"ADDREPLICA" } }' \
   "$SOLR_LOCAL_URL/solr/admin/autoscaling"

# node lost trigger
curl --header "Content-Type: application/json" \
   --request POST \
   --data '{ "set-trigger": { "name": "node_lost_trigger", "event": "nodeLost", 
"waitFor": "10s", "enabled": "true", "preferredOperation": "DELETENODE" } }' \
   "$SOLR_LOCAL_URL/solr/admin/autoscaling"

activeCollection=products
shards=2
replicationFactor=2
maxShardsPerNode=1

curl 
"$SOLR_LOCAL_URL/solr/admin/collections?action=CREATE&name=${activeCollection}&numShards=${shards}&replicationFactor=${replicationFactor}&collection.configName=products&wt=json&maxShardsPerNode=${maxShardsPerNode}"
 # &autoAddReplicas=true

# needed to not get into add-replica loop: 
https://lucene.472066.n3.nabble.com/Autoscaling-using-triggers-to-create-new-replicas-td4415260.html
curl -X POST -H 'Content-Type: application/json' \
"${SOLR_LOCAL_URL}/api/cluster/autoscaling" --data-binary '{
"set-cluster-policy": [
{"replica": "<2", "shard": "#ANY", "node": "#ANY"},
{"replica": "#EQUAL", "shard": "#ANY", "node": "#ANY"}
]
}'

Best regards,
Christian


Re: Phrase search and WordDelimiterGraphFilter not working as expected with mixed delimited and non-delimited tokens

2020-02-19 Thread Michael Gibney
There are many layers to this, but for the config you posted (applying
index-time WDGF configured to both split and catentate tokens), the
fundamental issue is that Lucene doesn't index positionLength, so the
graph structure (and token adjacency information) of the token stream
is lost when it's serialized to the index. Once the positionLength
information is discarded, it's impossible to restore/leverage it at
query time.

For now, if you use WGDF (or any analysis component capable of
generating "graph"-type output) at index-time, you'll have issues
unless you configure it such that it won't in practice generate graph
output. For WGDF this would mean either catenate output, or split
output, but not both on a single analysis chain. If you need both, one
option would be to index to (and search on) two fields: one for
catentated analysis, one for split analysis.

Graph output *is* respected at query-time, so you have more options
configuring WGDF on a query-time analyzer. But in that case, it's
worth being aware of the potential for exponential query expansion
(see discussion at https://issues.apache.org/jira/browse/SOLR-13336,
which restores a safety valve for extreme instances of this case).

Some other potentially relevant issues/links:
https://issues.apache.org/jira/browse/LUCENE-4312
https://issues.apache.org/jira/browse/LUCENE-7398
https://www.elastic.co/blog/multitoken-synonyms-and-graph-queries-in-elasticsearch
(Lucene, so applies also to Solr)
https://michaelgibney.net/lucene/graph/

On Wed, Feb 19, 2020 at 10:27 AM Jeroen Steggink | knowsy
 wrote:
>
> Hi,
>
> I have a question regarding phrase search in combination with a
> WordDelimiterGraphFilter (Solr 8.4.1).
>
> Whenever I try to search using a phrase where token combination consists
> of delimited and non-delimited tokens, I don't get any matches.
>
> This is the configuration:
>
> 
>
>  
>  
>generateWordParts="1"
>  generateNumberParts="1"
>  catenateWords="1"
>  catenateNumbers="0"
>  catenateAll="0"
>  splitOnCaseChange="1"
>  preserveOriginal="1"/>
>  
>  
>
>
>  
>  
>  
>
> 
>
>  omitTermFreqAndPositions="false" />
>
>
> Example document:
>
> {
>id: '1',
>text: 'mr. i.n.i.t. firstsirname secondsirname'
> }
>
> Queries and results:
>
> Query:
> "mr. i.n.i.t. firstsirname"
> -
> No result
>
> Query:
> "mr. i.n.i.t."
> -
> Result
>
> Query:
> "mr. i n i t"
> -
> Result
>
> Query:
> "mr. init"
> -
> Result
>
> Query:
> "mr init"
> -
> Result
>
> Query:
> "i.n.i.t. firstsirname"
> -
> No result
>
> Query:
> "init firstsirname"
> -
> No result
>
> Query:
> "i.n.i.t. firstsirname secondsirname"
> -
> No result
>
> Query:
> "init firstsirname secondsirname"
> -
> No result
>
>
> I don't quite understand why this is. When looking at the results of the
> analyzers I don't understand why it's working with just delimited or
> non-delimited tokens. However, as soon as the mixed combination of
> delimited and non-delimited is searched, there is no match.
>
> Could someone explain? And is there a solution to make it work?
>
> Best regards,
>
> Jeroen
>
>


Unable to start solr server on "Ubuntu 18.04 bash shell on Windows 10"

2020-02-19 Thread Prabhat Awasthi
Hello,

I am using Linux bash sell (Ubuntu app) on Windows 10 to run Solr on Ubuntu
18.04.

$ lsb_release -a
No LSB modules are available.
Distributor ID: Ubuntu
Description:Ubuntu 18.04.2 LTS
Release:18.04
Codename:   bionic

I already installed Java8 (Openjdk) on my Ubuntu environment.

$ java -version
openjdk version "1.8.0_242"
OpenJDK Runtime Environment (build 1.8.0_242-8u242-b08-0ubuntu3~18.04-b08)
OpenJDK 64-Bit Server VM (build 25.242-b08, mixed mode)

But I face error when I try to start SolrCloud on my Ubuntu system.
Could you please help to give me some pointers if I miss anything here ?
Please find below the full logs.

Thanks in advance.
- Prabhat

---
$ bin/solr start -e cloud
*** [WARN] *** Your open file limit is currently 1024.
 It should be set to 65000 to avoid operational disruption.
 If you no longer wish to see this warning, set SOLR_ULIMIT_CHECKS to false
in your profile or solr.in.sh
*** [WARN] ***  Your Max Processes Limit is currently 7823.
 It should be set to 65000 to avoid operational disruption.
 If you no longer wish to see this warning, set SOLR_ULIMIT_CHECKS to false
in your profile or solr.in.sh

Welcome to the SolrCloud example!

This interactive session will help you launch a SolrCloud cluster on your
local workstation.
To begin, how many Solr nodes would you like to run in your local cluster?
(specify 1-4 nodes) [2]:

Ok, let's start up 2 Solr nodes for your example SolrCloud cluster.
Please enter the port for node1 [8983]:

Please enter the port for node2 [7574]:

Solr home directory
/home/pawasthi/projects/solr_practice/ex1/solr-8.4.1/example/cloud/node1/solr
already exists.
/home/pawasthi/projects/solr_practice/ex1/solr-8.4.1/example/cloud/node2
already exists.

Starting up Solr on port 8983 using command:
"bin/solr" start -cloud -p 8983 -s "example/cloud/node1/solr"

*** [WARN] ***  Your Max Processes Limit is currently 7823.
 It should be set to 65000 to avoid operational disruption.
 If you no longer wish to see this warning, set SOLR_ULIMIT_CHECKS to false
in your profile or solr.in.sh
*Waiting up to 180 seconds to see Solr running on port 8983 [|]  bin/solr:
line 664:   293 Aborted (core dumped) nohup "$JAVA"
"${SOLR_START_OPTS[@]}" $SOLR_ADDL_ARGS -Dsolr.log.muteconsole
"-XX:OnOutOfMemoryError=$SOLR_TIP/bin/oom_solr.sh $SOLR_PORT
$SOLR_LOGS_DIR" -jar start.jar "${SOLR_JETTY_CONFIG[@]}"
$SOLR_JETTY_ADDL_CONFIG > "$SOLR_LOGS_DIR/solr-$SOLR_PORT-console.log" 2>&1*
 [|]  Still not seeing Solr listening on 8983 after 180 seconds!
tail: cannot open
'/home/pawasthi/projects/solr_practice/ex1/solr-8.4.1/example/cloud/node1/solr/../logs/solr.log'
for reading: No such file or directory

ERROR: Did not see Solr at http://localhost:8983/solr come online within 30
---


Re: Solr Upgrade socketTimeout issue in 8.2

2020-02-19 Thread kshitij tyagi
Hi Erick,

Thanks for the response.

1. We have started from zero documents and are observing socketTimeout
intermittently while indexing also.

2. We are not using schemaless as we are aware of the drawbacks and have a
well defined schema.

3. We have tried async but it does not help.

4. We tried increasing timeout as mentioned in
https://issues.apache.org/jira/browse/SOLR-13605 by using
withConnectionTimeout() and withSocketTimeout()

5. Our System was pretty stable earlier but as we upgraded to 8.2 from 6.6
version this issue is making Solrcloud system unstable

here is the detailed log which we are observing intermittently at time of
indexing:

ERROR (OverseerThreadFactory-9-thread-3-processing-n:prod-cloud2.net:8983_solr)
[   ] o.a.s.c.a.c.OverseerCollectionMessageHandler Error from shard:
http://prod-cloud2.net:8983/solr =>
org.apache.solr.client.solrj.SolrServerException: Timeout occurred while
waiting response from server at: http://prodcloud2.net:8983/solr
at
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:667)
org.apache.solr.client.solrj.SolrServerException: Timeout occurred while
waiting response from server at: http://prod-cloud2.net:8983/solr
at
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:667)
~[?:?]
at
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:262)
~[?:?]
at
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:245)
~[?:?]
at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1274)
~[?:?]
at
org.apache.solr.handler.component.HttpShardHandlerFactory$1.request(HttpShardHandlerFactory.java:176)
~[?:?]
at
org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:199)
~[?:?]
at
org.apache.solr.handler.component.HttpShardHandler.dt_access$927(HttpShardHandler.java)
~[?:?]
at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:1.8.0_131]
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
~[?:1.8.0_131]
at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:1.8.0_131]
at
com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.java:181)
~[metrics-core-4.0.5.jar:4.0.5]
at
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
~[?:?]
at
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.dt_access$304(ExecutorUtil.java)
~[?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
~[?:1.8.0_131]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
~[?:1.8.0_131]
at java.lang.Thread.run(Unknown Source) [?:1.8.0_131]
Caused by: java.net.SocketTimeoutException: Read timed out
at java.net.SocketInputStream.socketRead0(Native Method) ~[?:1.8.0_131]
at java.net.SocketInputStream.socketRead(Unknown Source) ~[?:1.8.0_131]
at java.net.SocketInputStream.read(Unknown Source) ~[?:1.8.0_131]
at java.net.SocketInputStream.read(Unknown Source) ~[?:1.8.0_131]
at
org.apache.http.impl.io.SessionInputBufferImpl.streamRead(SessionInputBufferImpl.java:137)
~[?:?]
at
org.apache.http.impl.io.SessionInputBufferImpl.fillBuffer(SessionInputBufferImpl.java:153)
~[?:?]
at
org.apache.http.impl.io.SessionInputBufferImpl.readLine(SessionInputBufferImpl.java:282)
~[?:?]
at
org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:138)
~[?:?]
at
org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:56)
~[?:?]
at
org.apache.http.impl.io.AbstractMessageParser.parse(AbstractMessageParser.java:259)
~[?:?]
at
org.apache.http.impl.DefaultBHttpClientConnection.receiveResponseHeader(DefaultBHttpClientConnection.java:163)
~[?:?]
at
org.apache.http.impl.conn.CPoolProxy.receiveResponseHeader(CPoolProxy.java:165)
~[?:?]
at
org.apache.http.protocol.HttpRequestExecutor.doReceiveResponse(HttpRequestExecutor.java:273)
~[?:?]
at
org.apache.http.protocol.HttpRequestExecutor.execute(HttpRequestExecutor.java:125)
~[?:?]
at
org.apache.solr.util.stats.InstrumentedHttpRequestExecutor.execute(InstrumentedHttpRequestExecutor.java:120)
~[?:?]
at
org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:272)
~[?:?]
at
org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:185)
~[?:?]
at org.apache.http.impl.execchain.RetryExec.execute(RetryExec.java:89)
~[?:?]
at
org.apache.http.impl.execchain.RedirectExec.execute(RedirectExec.java:110)
~[?:?]
at
org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:185)
~[?:?]
at
org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:83)
~[?:?]
at
org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:56)
~[?:?]
at
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:555)
~[?:?]
... 15 more

Thanks for the help, any further suggestions to drill down to problem?

Regards,
k