Hi,
Our indexer before starting does upload/reload of Solr configuration files
using ZK UPLOAD and RELOAD APIs. In this process zookeeper is not
stopped/restarted. ZK is alive and so are Solr nodes.
Doing this often causes following exception. Kindly note that the ZK
instance is standalone and not
I'm adding PatternReplaceCharFilterFactory to exclude characters. Looks like
this works.
-Original Message-
From: Tarala, Magesh
Sent: Thursday, July 30, 2015 10:37 AM
To: solr-user@lucene.apache.org
Subject: RE: StandardTokenizerFactory and WhitespaceTokenizerFactory
Using PatternRepl
Please be more specific as to why you think something is not working.
-- Jack Krupansky
On Thu, Jul 30, 2015 at 10:43 AM, Nitin Solanki
wrote:
> Hi,
>I am trying to index documents using solr cloud. After setting,
> to 6 ms in hard commit. Documents are visible instantly while
> ad
The reason is almost certainly because the query parser is splitting on
whitespace before the analysis chain gets the query - thus, each token
travels separately through your chain. Try it with quotes around it to
see if this is your issue.
Upayavira
On Thu, Jul 30, 2015, at 04:52 PM, Jack Schled
So, what you want is to duplicate a specific token, rename one of the
copies, and inject it with the same offset as the original. So GATE =>
gate, _gate but gate=>gate.
That, to me, is a custom token filter. You can probably use
KeywordRepeatFilterFactory as a base:
http://www.solr-start.com/javad
On 7/30/2015 10:46 AM, Robert Farrior wrote:
> We have a requirement to be able to have a master product catalog and to
> create a sub-catalog of products per user. This means I may have 10,000
> users who each create their own list of documents. This is a simple mapping
> of user to documents. The
On 7/30/2015 3:53 AM, Manohar Sripada wrote:
> We have Solr Cloud (version 4.7.2) setup on 64 shards spread across VMs.
> I see my queries to Solr taking exactly 50 sec intermittently (as
> someone said so :P). This happens once in 10 queries.
> I have enabled log level to TRACE on all the solr no
Thanks and I did think of the copy field option. So what you are suggesting
is that I have a copyfield in which I do not keep the lowercase factory
analyzer in my indexing/query chains.
I am afraid that would not help if my search query is complex with many
words (say a boolean with proximity oper
Hi,
We have a requirement to be able to have a master product catalog and to
create a sub-catalog of products per user. This means I may have 10,000
users who each create their own list of documents. This is a simple mapping
of user to documents. The full data about the documents would be in the m
Most probably because your solrconfig.xml is setting maxDocs for 1:
1. Then Solr will autoCommit EITHER with 1 document or
after has passed. Change your value in solrconfig.xml
to 1, don't forget to RELOAD the core, then test it again.
On Thu, Jul 30, 2015 at 12:13 PM, Nitin Solanki
wrote
Talha,
In your configuration, you have this set:
5
...which means it will consider the query "correctly spelled" and offer no
suggestions if there are 5 or more results. You could omit this parameter and
it will always suggest when possible.
Possibly, a better option would be to add "spellc
Hi,
I'm in the process of revising a schema for the search function of an
eCommerce platform. One of the sticking points is a particular use case of
searching for "xx yy" where xx is any number and yy is an abbreviation for
a unit of measurement (mm, cc, ml, in, etc.). The problem is that
search
Using PatternReplaceCharFilterFactory to replace comma, period, etc with space
or empty char will work?
-Original Message-
From: Tarala, Magesh
Sent: Thursday, July 30, 2015 10:08 AM
To: solr-user@lucene.apache.org
Subject: StandardTokenizerFactory and WhitespaceTokenizerFactory
I am in
Glad you find it useful Daniel!
Yeah its all driven from the browser. Splainer doesn't have a backend, its
just a bunch of html and javascript hosted on s3. So no worries about your
data being shared around.
It seems another common trend is just running it locally. I correspond with
quite a few f
Hi Edwards,
I am only sending 1 document for indexing then why it is
committing instantly. I gave to 6.
On Thu, Jul 30, 2015 at 8:26 PM Edward Ribeiro
wrote:
> Your is set to 1. This is the number of pending docs before
> autocommit is triggered too. You should set it to
I am indexing text that contains part numbers in various formats that contain
hypens/dashes, and a few other special characters.
Here's the problem: If I use StandardTokenizerFactory, the hypens, etc are
stripped and so I cannot search by the part number 222-333-. I can only
search for 222
Have you tried copyField with different field type for different
fields yet? That would be my first step. Make the copied field
indexed-only, not stored for efficiency.
And you can then either search against that copied field directly or
use eDisMax against both fields and give that field a higher
Your is set to 1. This is the number of pending docs before
autocommit is triggered too. You should set it to a higher value like
1, for example.
Edward
Em 30/07/2015 11:43, "Nitin Solanki" escreveu:
> Hi,
>I am trying to index documents using solr cloud. After setting,
> to 6
Hi,
I am trying to index documents using solr cloud. After setting,
to 6 ms in hard commit. Documents are visible instantly while
adding them. Not commiting after 6 ms.
I have added Solr log. Please check it. I am not getting exactly what is
happening.
*CURL to commit documents:*
Hi,
I need the capability to search for /GATE/ separately from /gate/.
I cannot remove the lowercase filter factory in both my search and analysis
chains since that will break many other search scenarios.
Is there a way to payload/mark an ALL CAPS word in the index analyzer chain
before it gets
"Although I'm not sure why you took this approach instead of
supporting simple built-in basic auth and let us configure security
the "old/easy" way"
Going with Jetty basic auth is not useful in a large enough cluster.
Where do you store the credentials and how would you propagate it
across the c
Hi Mohsen,
this is the perfect place for the *positionIncrementGap *attribute for your
field type*.*
First of all when phrase or positional searches are necessary you need to
store term positions in your index.
The position increment gap will increment the position when a multi valued
field happ
Hi Nutch,
are you sure you are using the proper parameters ?
I can not see the "highlight" param in the suggester configuration!
>From the issue you linked, it seems it is necessary to disable highlighting
( default =true) .
I see it as query param for the /suggest search handler.
Am I wrong or yo
Hi Ashish, are we talking about Analysis at query or Index time or both ?
As Erick say I found really hard to believe for this combination in a
classic search.
Are you trying to provide something "special" ?
Ngram token filter will produce a setof ngram out of your token:
token
to ok ke en in
Hi,
We have Solr Cloud (version 4.7.2) setup on 64 shards spread across VMs. I
see my queries to Solr taking exactly 50 sec intermittently (as someone
said so :P). This happens once in 10 queries.
I have enabled log level to TRACE on all the solr nodes. I didn't find any
issue with the query time
I am still experiencing https://issues.apache.org/jira/browse/SOLR-6648 issue
with solr 5.2.1.
even if i send highlight=false solr returns me highlighted suggestions. Any
idea why this is happening?
My configurations :
*URL :
*http://solrhost:solrpost/mycorename/suggest?suggest.dictionary=altSu
How can I recognize line breaks and do not allow matching of a quoted query
in the following example.
I have two documents with just one text field:
1. AAA BBB
CCC DDD
2. BBB CCC
DDD AAA
User enters query "BBB CCC". How can I configure tokenizers so that Solr
only returns doc #2?
Thanks,
Mohs
27 matches
Mail list logo