The field will contain semantic information about the document. It would
need to be searchable plus it will contain information that will be used as
part of the score. Probably a payload that will be used by a custom scorer.
On Friday, August 2, 2013, Mikhail Khludnev wrote:
> Marc,
>
> I wonder
Hey guys,
It is by no means perfect or pretty, but I use this script below to
build Solr into a .deb package that installs Solr to /opt/solr-VERSION
with 'example' and 'docs' removed, and a symlink to /opt/solr. When
building, the script wget's the tgz, builds it in a tmpdir within the
cwd an
Hi Dmitry,
Thanks, It was a toothing problem, fixed now, please try the fresh checkout
AND add the following to your arguments: -t /solr/core1
that sets the path under which solr should be contacted, the handler is set
in the jmeter configuration, so if you were using different query handlers
tha
On 8/2/2013 4:44 PM, Robert Zotter wrote:
> When was this feature added?
>
> This did indeed fix the problem.
>From what I can tell, the ability to set autoGeneratePhraseQueries was
added in 3.4 or 3.5.
Thanks,
Shawn
You seem to be mixing a couple of different concepts here. "Prospective
search" or reverse search, (sometimes called alerts) is a logistics matter,
but how to match terms is completely different.
Solr does not have the exact "percolate" feature of ES, but your examples
don't indicate a need fo
We have a set number of known terms we want to match against.
In Index:
"term one"
"term two"
"term three"
I know how to match all terms of a user query against the index but we would
like to know how/if we can match a user's query against all the terms in the
index?
Search Queries:
"my search
Hi Erick,
Yes we have lots of documents with the same score. We turned off IDF and on
some fields also the norms and frequencies. And so on a business point of
view we need to shuffle those results in order to give the opportunity to
any document with the highest score to be in first.
The problem
When was this feature added?
This did indeed fix the problem.
On Aug 2, 2013, at 3:27 PM, Shawn Heisey wrote:
> On 8/2/2013 4:16 PM, Robert Zotter wrote:
>> The problem is the query get's expanded to "1 Foo" not ( "1" OR "Foo")
>>
>> 1Foo
>> 1Foo
>> +DisjunctionMaxQuery((name_textsv:"1 foo"))
Hi Kuro,
It is possible with https://issues.apache.org/jira/browse/SOLR-1604 . But this
is a sold-plugin.
There is built-in http://wiki.apache.org/solr/SurroundQueryParser, but syntax
is a little bit different. q = san w fran*
From: T. Kuro Kurosaka
To: so
On 8/2/2013 4:16 PM, Robert Zotter wrote:
The problem is the query get's expanded to "1 Foo" not ( "1" OR "Foo")
1Foo
1Foo
+DisjunctionMaxQuery((name_textsv:"1 foo")) ()
+(name_textsv:"1 foo") ()
DisMaxQParser
This looks like you have autoGeneratePhraseQueries turned on for the
field defin
The problem is the query get's expanded to "1 Foo" not ( "1" OR "Foo")
1Foo
1Foo
+DisjunctionMaxQuery((name_textsv:"1 foo")) ()
+(name_textsv:"1 foo") ()
DisMaxQParser
On Aug 2, 2013, at 1:55 PM, "Jack Krupansky" wrote:
> Is your q.op "AND" or "OR"?
>
> Add debugQuery=true to your query req
I'm using field collapsing to group documents by a single field and have run
into something unexpected with how sorting of the groups works. Right now I
have each group return one document. The documents within each group are sorted
by a field ("price") in ascending order using group.sort so tha
I am trying to get autocorrect and suggest feature work on my solr 4.4
setup.
As recommended here: http://wiki.apache.org/solr/Suggester, this is my
solrconfig: http://apaste.info/eBPr
Where "allText" is a copy field which indexes all the content I have in
document title, description etc.
I am t
Is your q.op "AND" or "OR"?
Add debugQuery=true to your query request and look at the parsed query
string to see what query is actually generated.
-- Jack Krupansky
-Original Message-
From: Robert Zotter
Sent: Friday, August 02, 2013 4:21 PM
To: solr-user@lucene.apache.org
Subject:
bq: We sort first by the score and then by a random field.
This is highly unlikely to produce anything useful. Your
secondary sort (random field) will only be used if your
primary sort (score) matches exactly, which is usually
quite rare. Unless you're somehow doing a constant-score
query (e.g. *:
Would this question be more appropriate on Solr-Dev?
Thank you in advance,
O. O.
O. Olson wrote
> Hi,
>
> I am using Solr with the VelocityResponseWriter.
> http://wiki.apache.org/solr/VelocityResponseWriter I am wondering if
> there is anyway to add my own Utility Class i.e. how do I pu
I'm trying to match such terms as "1Foo" agains "Foo" but I can't seem to get
it working. I'm using the WordDelimiterFactory and everything looks good in the
/admin/analysis.jsp but when ever I do a "real" query, It doesn't return any
results? Any thoughts on why this may be? Thanks
Marc,
I wonder what's type of the field what kind of search you need on it
filtering/ranking/boosting etc.
Thanks
On Fri, Aug 2, 2013 at 11:00 PM, Marc Brette wrote:
> This is something I am considering.
>
> Ideally, I'd like to use the same index though.
> I do need to query with other const
This is something I am considering.
Ideally, I'd like to use the same index though.
I do need to query with other constraint but that could be resolved to some
extent by merging results post query.
The real headache with different indexes is management: deleting document,
backup/restore. We also h
Marc,
Do you need to be able to query this field at the same time as other
fields, or is the searching case isolated?
Because if you can isolate searches that hit this field to just this field,
you could do it with a sidecar index and joins.
Michael Della Bitta
Applications Developer
o: +1 646
Unfortunately, it needs to be searchable
Very good pointer anyway, I'll keep that in mind
On Friday, August 2, 2013, Michael Della Bitta wrote:
> Hi Marc,
>
> Have you considered using ExternalFileField for this?
> On Aug 2, 2013 11:54 AM, "Marc Brette" >
> wrote:
>
> > Hi,
> > I would like to
Or shingles, presuming you want to tokenize and output unigrams.
On Aug 2, 2013, at 11:33 AM, Walter Underwood wrote:
> Search against a field using edge N-grams. --wunder
>
> On Aug 2, 2013, at 11:16 AM, T. Kuro Kurosaka wrote:
>
>> Is there a query parser that supports a phrase query with p
Search against a field using edge N-grams. --wunder
On Aug 2, 2013, at 11:16 AM, T. Kuro Kurosaka wrote:
> Is there a query parser that supports a phrase query with prefix query at the
> end, such as "San Fran*" ?
>
> --
> -
> T. "Kuro" Kurosaka • Senio
Is there a query parser that supports a phrase query with prefix query
at the end, such as "San Fran*" ?
--
-
T. "Kuro" Kurosaka • Senior Software Engineer
Thanks, Ali,
That is fine, I could do that (probably even by calling some serlvet end
point similar to unload core functionality in the admin). But another
question that arises is how to provision the new cores without restarting
the solr instance?
On Fri, Aug 2, 2013 at 7:41 PM, Ali, Saqib wro
On 8/2/2013 11:33 AM, SolrLover wrote:
Is PositionFilter is deprecated as of Lucene 4.4? Is there any other
alternate way to implement that functionality?
The javadoc says it's deprecated in 4.4, and mentions a couple of ways
around it. The first recommendation, setting whether or not to
aut
Is PositionFilter is deprecated as of Lucene 4.4? Is there any other
alternate way to implement that functionality?
--
View this message in context:
http://lucene.472066.n3.nabble.com/PositionFilter-is-deprecated-as-of-Lucene-4-4-tp4082245.html
Sent from the Solr - User mailing list archive at
I am trying to match the keywords with / without white space but one of the
case fails always..
For ex:
I am indexing 4 documents
name: wal mart
name: walmart
name: WalMart
name: Walmart
Now searching on name either using
wal mart
walmart
Walmart
WalMart
should return all the above 4 documents
On 8/2/2013 8:13 AM, Anca Kopetz wrote:
Then we optimized the index to 1 segment / 0 deleted docs and we got
+40% of QPS compared to the previous test.
Therefore we thought of optimizing the index every two hours, as our
index is evolving due to frequent commits (every 30 minutes) and thus
the p
On 8/2/2013 2:48 AM, Furkan KAMACI wrote:
I use Solr 4.2.1 as SolrCloud. My live indexes will be search by huge
amounts of users and I don't want to have anything wrong. I have some
criteria for my indexes. i.e. there mustn't be spam documents at my index
(I have a spam detector tool), some docum
You must have some other relevancy rules, right? Otherwise, for every 10
results you wish to display, just take the top 7 pizza results, then the
first 2 bread results, and then the top healthy result. But... I suspect
that this would not be acceptable because... you have additional rules that
Hi all.
In my recipes website I want to create a page where users can browse and
the return must be based on this rule:
- 70% of the results must be pizzas (type:pizza)
- 20% of the results must be breads (type:bread)
- 10% of the results must be healthy (type:healthy)
But I don't know how to qu
Dmitry,
That is expected behaviour. You need to manually remove the original core.
Thanks.
On Fri, Aug 2, 2013 at 6:03 AM, Dmitry Kan wrote:
> Hello list,
>
> I was wondering, if what I see with the split shard a correct behaviour or
> is something wrong.
>
> Following this article:
>
> http:
On 8/2/2013 6:47 AM, Brian Hurt wrote:
I seem to recall somewhere in the documention that the create function on
non-cloud solr doesn't copy the config files in, you have to copy them in
by hand. Is this correct? If so, can anyone point me to where in the docs
it says this, and if there are any
Hello,
We are trying to make a random sort with a query where we are grouping
results by a field. On one example we always have one of the results on the
top of the query result. We sort first by the score and then by a random
field.
We think that is because this group has 214 items inside and th
Hi Marc,
Have you considered using ExternalFileField for this?
On Aug 2, 2013 11:54 AM, "Marc Brette" wrote:
> Hi,
> I would like to completely populate a field for all the documents in the
> index, without re-indexing the documents.
>
> I know Solr supports 'Atomic Update', but this is no a re
Hi,
I would like to completely populate a field for all the documents in the
index, without re-indexing the documents.
I know Solr supports 'Atomic Update', but this is no a real incremental
update of a document: it costs as much as re-indexing the document (and
require to store the document).
As
> I tried below two methods(edismax & OR) in solrj java coding but I did not
> get the results in order, like listing matching records at the top. Am I
> doing anything wrong in below java coding?
>
>
> Method 1
> :
> SolrQuery query = new SolrQuery().setStart(first).setRows(
>
"product promotions" = "query elevation"
See:
http://wiki.apache.org/solr/QueryElevationComponent
https://cwiki.apache.org/confluence/display/solr/The+Query+Elevation+Component
Or, boost the query using a function query referencing an external file
field that gets updated for promotions.
--
Yeah, roughly.
The example I added to the book (for EAR#5) is generalized and takes a field
name parameter, a regular expression, to match, and a case sensitivity flag.
And works for multivalued fields. And optionally logs docs that are skipped
-- Jack Krupansky
-Original Message-
F
Hi,
We are trying to improve the performance of our Solr Search application in
terms of QPS (queries per second).
We tuned SOLR settings (e.g. mergeFactor=3), launched several benchmarks and
had better performance results, but still unsatisfactory for our traffic volume.
Then we optimized the
Jack, thanks for the response. So, adding something as simple as the
following to the processAdd() function should do the trick in your opinion?
this_title = doc.getFieldValue("title");
if (this_title == "Slide 1"){
return false;
}
Regards,
ADS
--
View
Ok, i managed to load a config file from the node's sharedLib directory by
pointing it to `../../lib/file`. This works fine in normal mode, in cloud mode
all it does is attempting to find it in Zookeeper by trying
/configs/COLLECTION_NAME/path. Properties such as ${solr.home} are not
recognized
Hi Erick
Many Thanks for your reply. I got your point. one question on this: is it
possible to give more priority to those docs which has higher indexing time
boosting against query time boosting. I am trying to achieve product
promotions using this implementation. can you please guide how should
It appears that this is simpler than I thought: in SOLR 4.4, at least,
there is a dataSource class named "FieldStreamDataSource" that I can use
directly with the TikaEntityProcessor. Given a blob column named DOCIMAGE,
I can use the following Tika entity:
...
Hello list,
I was wondering, if what I see with the split shard a correct behaviour or
is something wrong.
Following this article:
http://searchhub.org/2013/06/19/shard-splitting-in-solrcloud/
I have issued a low-level core split query:
http://localhost:8982/solr/admin/cores?core=core1&action=
Im using SolrCloud with two solr instances in two separate tomcats and
Zookeeper.
I have created a collection with 4 shards ( 2 in each solr) and was able to
distribute data.
Now I brought down solr2, then searched.
1. As expected solr2 was down. But solr1 was throwing
*no servers hosting shard
I seem to recall somewhere in the documention that the create function on
non-cloud solr doesn't copy the config files in, you have to copy them in
by hand. Is this correct? If so, can anyone point me to where in the docs
it says this, and if there are any plans to change this? Thanks.
Brian
Yes, all the usual config files are well under 1MB and work as expected. This
file is under 2MB and the limit i set is 5MB. Setting jute.maxbuffer (all
lowercase) did work during a test a long time ago but we'd like to put the new
features in production and we're stuck at this trivial issue :)
Hmmm, does it work with smaller config files? There's been a limit of 1M
for ZK files, and I'm wondering if your setup would work with, say, 2M
configs as a check that it's "something else" rather than just the 1M limit.
FWIW,
Erick
On Fri, Aug 2, 2013 at 8:18 AM, Markus Jelsma wrote:
> Hi,
>
>
Hi,
I have a few very large configuration files but it doens't work in cloud mode
due to the KeeperException$ConnectionLossException. All 10 Solr nodes run trunk
and have jute.maxbuffer set to 5242880 (5MB). I can confirm it is set properly
by looking at the args in the Solr GUI. All Zookeepers
Mark,
thank you for your reply.
> You have to have everything setup right, but yes, I think this should work.
> You would probably want to use a load balancer for failover or something, but
> I don't know what would stop the scatter gather from working under the right
> conditions.
Yes, the sch
Two solutions:
1. An update processor that copies the float field to an int field and
rounds it. you can do that with a JavaScript script using the
StatelessScriptUpdate processor. Then query on the integer field
2.Do range queries: xyz:[2.0 TO 2.5]
-- Jack Krupansky
-Original Message--
Add &debug=all to your query, that'll show you exactly how the scores
are calculated. But the most obvious thing is that you're boosting
on the titleName field in your query, which for doc 123 does NOT
contain "phone" so I suspect the fact that "phone" is in the titleName
field for 122 is overridin
Well, this is probably not a rogue query. You can test this, of course
by replaying all your queries on a test system. My guess is that it's
just too much stuff on too small a box.
Or you could have poorly configured Solr parameters. I've seen, for
instance, the filterCache sized at 1M. Which runs
Sounds like you should have separate clusters. Maybe "production" vs.
"staged". Build up your new document base in "staged", test it all out and
then "flip a switch" and staged goes live and then the old production can be
recycled.
Or, maybe have a "pipeline" where new docs are indexed on put
Hello All,
I want to boost certain products on particular keywords. for this i am
using solr's indexing time boosting feature. i have given index time
boosting with "1.0" value to all documents in my solr indices. now what i
am doing is when user want to boost certain product i just increase index
I think that form of routing is optional, so it's not required for Cloud per
se.
Still, I think we're back to what I suggested, that this is another Solr
feature that CAN be used, but doesn't necessarily work everywhere in Solr
(like field names containing white space or special characters.)
Then you need to do two things:
1> index the values as integers, i.e. do the rounding at input time.
You could use a custom update element to do this or (simpler)
just do it in your indexing step. Using the round methods in your
SQL statement has already been mentioned.
2> round the val
I have tried in 4.4 too. It also produces same kind of problem only.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Split-Shard-Document-loss-and-down-time-tp4082002p4082180.html
Sent from the Solr - User mailing list archive at Nabble.com.
Solr parameters listed in dashboard:
-DzkHost=localhost:2181,172.27.5.121:2181,172.27.5.122:2181
-XX:+UseConcMarkSweepGC
-Xmx3072m
-Djava.awt.headless=true
Mem usage in last days: http://i42.tinypic.com/29z5rew.png
It's production system and there's too many requests to detect which query
is bugg
Another option would be to write a SolrJ program that runs
inside the firewall protecting the DB and writes to the Solr
server.
FWIW,
Erick
On Thu, Aug 1, 2013 at 8:16 PM, Jack Krupansky wrote:
> Consider running Apache ManifoldCF local to the database and then remote
> to Solr.
>
> Or, export
Been there, done that .
On Thu, Aug 1, 2013 at 4:46 PM, Luís Portela Afonso
wrote:
> Oh my god. Thanks for notice. The field name its wrong. It should be
> enclosure_type. I'm so sorry.
>
> On Aug 1, 2013, at 6:33 PM, Jack Krupansky
> wrote:
>
> > Are you sure the “enclosure_title” field is pop
On Fri, Aug 2, 2013 at 3:18 PM, Furkan KAMACI wrote:
> I don't want them to be visible at live index
> until I remove them. I should make some verification related to Solr. i.
>
assuming you comment proposal about deleteQuery.
once again, you can add all documents, then before commit, send few
de
You should be able to mine the logs for the query, as:
webapp=/solr path=/select params={q=manu:samsung} hits=1 status=0 QTime=3
Admittedly you have to do some reconstruction, but it's there. From there,
you should
be able to run the query against a warmed Solr (but not one that has had
that parti
(13/08/02 17:53), Jochen Lienhard wrote:
Hello,
we have a problem with some special characters: for example æ
We are using the ICUTranformFilterFactory for indexing and searching.
We have some documents with "urianae" and with "urianæ"
If I search "urainae" so I find only the versions with "
Hi Roman,
Sure:
python solrjmeter.py -a -x ./jmx/SolrQueryTest.jmx -q
/home/dmitry/projects/lab/solrjmeter/queries/demo/demo.queries -s localhost
-p 8983 -a --durationInSecs 60 -R test
This is vanilla install (git clone) except for one change that I had to do
related to solr cores:
> git diff
d
And last I knew (admittedly a long time ago) Query Elevation Component
would fail with a non-string type. So in Robert's case things would run
along fine forever until using QEV... Which, to be fair, may be never ..
This may have changed, but is an example of how the not
being a string type can p
>From the Wiki page:
4.3 however due to bugs found after 4.3 release, it is recommended that
you wait for release 4.3.1 before using this feature
So it would be great if you could try it with a newer Solr (4.4 was
recently released) because if it's still a problem we need to know.
Best
Erick
Thanks for the feedback. I don't want them to be visible at live index
until I remove them. I should make some verification related to Solr. i.
2013/8/2 Mikhail Khludnev
> I don't think you get it for free in Solr (especially in SolrCloud).
>
> You can experiment with own handler which does Luce
Roman:
Thanks for putting this together. I confess I haven't dug in in detail yet,
but having the numbers available is a nice resource.
Best
Erick
On Thu, Aug 1, 2013 at 6:23 PM, Roman Chyla wrote:
> On Thu, Aug 1, 2013 at 6:11 PM, Shawn Heisey wrote:
>
> > On 8/1/2013 2:08 PM, Roman Chyla w
Thanks a lot for both replies. Helped me a lot. It seems that
EdgeNGramFilterFactory on query analyzer was really my problem, I'll have to
test it a little more to be sure.
As for the "bf" parameter, I thinks it's quite fine as it is, from
documentation:
"the bf parameter actually takes a list o
Hi.
What is the version of Solr?
How exactly have you deployed it?
Have you used the .war file from the dist/ folder?
When do you see those messages?
Best regards,
Artem Karpenko.
02.08.2013 6:24, anpm1989 пишет:
Hello everyone,
I am a newbie with Solr and J2E server. I trying to deploy Solr
I don't think you get it for free in Solr (especially in SolrCloud).
You can experiment with own handler which does Lucene NRT
http://blog.mikemccandless.com/2011/06/lucenes-near-real-time-search-is-fast.htmlit
can open reader w/o commit that causes flushing rambuffer on disk.
Other approach, is
Hello,
we have a problem with some special characters: for example æ
We are using the ICUTranformFilterFactory for indexing and searching.
We have some documents with "urianae" and with "urianæ"
If I search "urainae" so I find only the versions with "urianae" but not
the "urianæ"
Only if I
I use Solr 4.2.1 as SolrCloud. My live indexes will be search by huge
amounts of users and I don't want to have anything wrong. I have some
criteria for my indexes. i.e. there mustn't be spam documents at my index
(I have a spam detector tool), some documents should be at first result
page (or with
Your solution worked perfectly in solr!. You have any idea how to achieve the
same with solrj java code?.
I tried below two methods(edismax & OR) in solrj java coding but I did not
get the results in order, like listing matching records at the top. Am I
doing anything wrong in below java coding?
77 matches
Mail list logo