separate fields, not in the "lat,lon" format.
So far, all my attempts to use multivalued failed, and I would greatly
appreciate some help.
Thanks!
Chris
(Resending because DMARC-compliant ESPs bounced the previous version)
�
I'm looking for a way to do an atomic update, but if a certain
condition exists on the existing document, abort the update.
�
Each document has the fields id, count, and value.� The source data ha
The docs for the JSON facet API tell us that the default ranges are inclusive
of the lower bounds and exclusive of the upper bounds. �I'd like to do the
opposite (exclusive lower, inclusive upper), but I can't figure out how to
combine the 'include' parameters to make it
work.
�
�
facturer
which "works" on a test
subset of 1,000 manufacturers.� However, there are two problems:
1) This query returns all the manufacturers, so I have to iterate over the
entire response object to extract the ones I want.
2) The query on the whole data set takes more than 600 seconds to return, which
doesn't fit
our target response time
�
How can I perform this query?
We're using solr version 5.5.5.
�
�
Thanks,
Chris
�
;:{},
� � "facet_intervals":{},
� � "facet_heatmaps":{}},
� "facets":{
� � "count":0}}
�
�
> Hi Chris,
> You mention it returns all manufacturers? Even after you apply filters
> (don’t see filter in
I'm not looking for products where the price is in the range [100, 200].
I'm looking for manufacturers for which the sum of the prices of all of their
products is in the range [100, 200].
�
> Hi Chris,
>
> I assumed that you apply some sort of fq=price:[100 TO 200] to
Emir,
It certainly seems like I'll need to use streaming expressions.
Thanks for your help!
Chris
> Hi Chris,
> I misunderstood your requirement. I am not aware of some facet result
> filtering feature. What you could do is sort facet results by sum and load
> page by p
I'm trying to figure out how to structure this query.
I have two types of documents: items and sources. Previously, they were all in
the same collection. I'm now testing a cluster with separate collections.
The items collection has 38,034,895,527 documents, and the sources collection
has 41
ments have 65 fields)
Thank you for your help.
�
Chris
�
Original Message
Subject: Re: joining across sharded collection
From: "Erick Erickson"
Date: Sat, December 9, 2017 10:1
have the range that i need, so if i can find the file, maybe
i can change it manually.
My next question is - can we have a catch all for ranges, i mean if things
don't match any other range then insert in this shard..is this possible?
Kindly advice.
Chris
have the range that i need, so if i can find the file, maybe
i can change it manually.
My next question is - can we have a catch all for ranges, i mean if things
don't match any other range then insert in this shard..is this possible?
Kindly advice.
Chris
tionFactor=1<http://localhost:8983/solr/admin/collections?action=CREATE&name=mycollection&numShards=3&replicationFactor=4>
Please advice.
Regards,
Chris
On Tue, Oct 15, 2013 at 8:07 PM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:
> How did you create these s
ot;},
"core_node7":{
"state":"active",
"core":"web_shard2_replica2",
"node_name":"64.251.14.47:1984_solr",
"base_url":"http://64.251.14.47:1984/solr
at 4:12 PM, Chris wrote:
> Hi,
>
> Please find the clusterstate.json as below:
>
> I have created a dev environment on one of my servers so that you can see
> the issue live - http://64.251.14.47:1984/solr/
>
> Also, There seems to be something wrong in zookeeper, when we try
Also, is there any easy way upgrading to 4.5 without having to change most
of my plugins & configuration files?
On Wed, Oct 16, 2013 at 4:18 PM, Chris wrote:
> oops, the actual url is -http://64.251.14.47:1981/solr/
>
> Also, another issue that needs to be raised is the creation
oh great. Thanks Primoz.
is there any simple way to do the upgrade to 4.5 without having to change
my configurations? update a few jar files etc?
On Wed, Oct 16, 2013 at 4:58 PM, wrote:
> >>> Also, another issue that needs to be raised is the creation of cores
> from
> >>> the "core admin" sec
rs with the new ones
> (if needed - check release notes for version 4.4.0 and 4.5.0 where all the
> versions of external tools [tika, maven, etc.] are stated) and deploy the
> updated WAR file to servlet container.
>
> Primoz
>
>
>
>
> From: Chris
> To: sol
;:"compositeId"},
"replicationFactor":"1"},
I am using this to add -
CloudSolrServer solrCoreCloud = new
CloudSolrServer(cloudURL);
solrCoreCloud.setDefaultCollection("web");
UpdateResponse up = s
)
at java.lang.Thread.run(Thread.java:722)
On Thu, Oct 17, 2013 at 6:31 PM, Chris wrote:
> Wow thanks for all that, i just upgraded, linked my plugins & it seems
> fine so far, but i have run into another issue
>
> while adding a document to the so
page...?
Thanks,
Chris
1854)
Please help.
Chris
Hi All,
I get characters like -
�� - CTA -
in the solr index. I am adding Java beans to solr by the addBean() function.
This seems to be a character encoding issue. Any pointers on how to
resolve this one?
I have seen that this occurs mostly for japanese chinese ch
/c26alibaba/
All text in title comes like -
- �
-
�
Can you please advice?
Chris
On Tue, Oct 29, 2013 at 11:33 PM, Rajani Maski wrote:
>
encoding then those
characters appear as some japanese characters I think.
But then this solution doesn't work for all special characters as I can
still see some of them...isn't there an encoding that can cover all the
characters whatever they might be? Any ideas on what do i do?
Regard
or UTF-8, and misinterprets
> SHIF-JIS encoded characters.
>
> You need to debug to find out where the characters
> get corrupted.
>
>
> On 11/04/2013 11:15 PM, Chris wrote:
>
>> Sorry, was away a bit & hence the delay.
>>
>> I am inserting java strings in
Hi Gurus,
I have a relevancy ranking questrion -
1. I have fields - title, domain, domainrank in index.
2. I am looking to maybe load a txt file of prefered domains at solr
startup & boost documents from those domains if keyword matches text in
title or domain (if it exactly matches the domain,
Hi,
I have a document structure that looks like the below. I would like to
implement something like -
(urlKeywords:"+keyword+" AND domainRank:[3 TO 1] AND adultFlag:N)^60 " +
"OR (title:"+keyword+" AND domainRank:[3 TO 1] AND adultFlag:N)^20 " +
"OR (title:"+keyword+" AND domainRank:[1
Dear Varun,
Thank you for your replies, I managed to get point 1 & 2 done, but for the
boost query, I am unable to figure it out. Could you be kind enough to
point me to an example or maybe advice a bit more on that one?
Thanks for your help,
Chris
On Tue, Feb 4, 2014 at 3:14 PM, Varun Tha
th 40 Websites, set B with 7 websites)
Does it sound possible with SOLR?
Do I have to expect custom development? If so, how much?
Thank you in advance
Bye, Chris
Hello,
We currently have a ton of documents that we would like to index and
make search-able. I came across solr and it seems like it offers a lot
of nice features and would suite our needs.
The documents are in similar structure to java code, blocks
representing functions, variables, comment blo
Maybe you jetty need to turning
how many memory in your system ?
Can you show the processes information with the java processes ?
above
Chris
2008/6/2 Bram de Jong <[EMAIL PROTECTED]>:
> Hello all,
>
>
> Still running tests on solr using the example jetty cont
HTTP headers are case insensitive
Original message From: simon Date:
5/29/18 12:17 PM (GMT-05:00) To: solr-user
Subject: Re: CURL command problem on Solr
Could it be that the header should be 'Content-Type' (which is what I see
in the relevant RFC) rather than 'Content-
changed), or by creating a new set of
AMIs and then terminating the instances.Is there a better way to do this? I'm
not facing any real problems with this setup, but I want to make sure I'm not
missing something obvious.Thanks,Chris
:
: Indeed, it's a doc problem. A long time ago in a Solr far away, there
: was a bunch of effort to use the "default" collection (collection1).
: When that was changed, this documentation didn't get updated.
:
: We'll update it in a few, thanks for reporting!
Fixed on erick's behalf because he
: Not sure I'm onboard with the first proposed solution, but yes, I'd open a
: JIRA issue to discuss.
we should standardize the context keys to use use fully
qualified (org.apache.solr.*) java class name prefixes -- just like we do
with the logger names themselves.
:
: - Mark
:
: On Mon, Jan
: ((attr_search:8 attr_search:gb)~2^5.0)
:
: I hope to be right, but I expect to find a boost in both the values
: matches.
1) "boost" information should show up as a detail of the "queryWeight",
which is itself a detail of the "weight" of term clauses -- in the output
you've included below, y
: What I’m finding is that now and then base_url for the replica in
: state.json is set to the internal IP of the AWS node. i.e.:
:
: "base_url":"http://10.29.XXX.XX:8983/solr”,
:
: On other attempts it’s set to the public DNS name of the node:
:
: "base_url":"http://ec2_host:8983/solr”,
:
:
: Boost Query (bq) accepts lucene queries. E.g. bq=price:[50 TO 100]^100
: boost and bf parameters accept Function queries, e.g. boost=log(popularity)
while these statements are both true, they don't tell the full story.
for example you can also specify a function as a query using the
appropri
: Thanks, very helpful. I think I'm on the right track now, but when I do a
: post now and my UpdateRequestProcessor extension tries to add a field to a
: document, I get:
:
: RequestHandlerBase org.apache.solr.common.SolrException: ERROR: [doc=1]
: Error adding field 'myField'='2234543'
:
: Th
Concrete details are crucial -- what exactly are you trying, what results
are you getting, how do those results differ from what you expect?
https://wiki.apache.org/solr/UsingMailingLists
Normally, even when someone only gives a small subset of the crucial
details needed to answer thei
doing queries, each user would only ever see 2 documents, Document 1
, plus whichever other document they specifically owned.
If there are web pages or book chapters I can read or re-read that address
this class of problem, those references would be great.
-Chris.
: I'm testing this on Windows, so that maybe a factor too (the OS is not
: releasing file handles?!)
specifically: Windows won't let Solr delete files on disk that have open
file handles...
https://wiki.apache.org/solr/FAQ#Why_doesn.27t_my_index_directory_get_smaller_.28immediately.29_when_i_de
: I have a requirement where I need to retrieve 1 to 15000 records at a
: time from SOLR.
: With 20 or 100 records everything happens in milliseconds.
: When it goes to 1000, 1 it is taking more time... like even 30 seconds.
so far all you've really told us about your setup is that some
Solr is the popular, blazing fast, open source NoSQL search platform from
the Apache Lucene project. Its major features include powerful full-text
search, hit highlighting, faceted search, dynamic clustering, database
integration, rich document (e.g., Word, PDF) handling, and geospatial
sear
Sounds like a mismatch in the way the BooleanQuery explanation generation
code is handling situations where there is/isn't a coord factor involved
in computing the score itself. (the bug is almost certainly in the
"explain" code, since that is less rigorously tested in most cases, and
the sco
: We are using Solrj to query our solr server, and it works great.
: However, it uses the binary format wt=javabin, and now when I'm trying
: to get better debug output, I notice a problem with this. The thing is,
: I want to include the explain data for each search result, by adding
: "[expla
BCE dates have historically been problematic because of ambiguity in both
the ISO format that we use for representing dates as well as the internal
java representation, more details...
https://issues.apache.org/jira/browse/SOLR-1899
..the best work around I can suggest is to use simple numeri
: What I'm wondering is, what should one do to fix this issue when it
: happens. Is there a way to recover? after the WARN appears.
It's just a warning that you have a sub-optimal situation from a
performance standpoint -- either committing too fast, or warming too much.
It's not a failure, a
: We do POST to add data to Solr v4.7 and Solr v5.3.2 respectively. The
: attachmentnames are in 789, 456, 123 sequence:
...
: And we do GET to select data from solr v4.7 and solr v5.3.2 respectively:
: http://host:port/solr/collection1/select?q=id:1&wt=json&indent=true
...
: Is
403 means "forbidden"
Something about the request Solr is sending -- or soemthing about the IP
address Solr is connecting from when talking to medium.com -- is causing
hte medium.com web server to reject the request.
This is something that servers may choose to do if they detect (via
headers
:
:
1) as a general rule, if you have a delcaration which includes
"WEB-INF" you are probably doing something wrong.
Maybe not in this case -- maybe "search-webapp/target" is a completley
distinct java application and you are just re-using it's jars. But 9
times out of 10, when people have
1) The hard coded implicit default sort order is "score desc"
2) Whenever a sort results in ties, the final ordering of tied documents
is non-deterministic
3) currently the behavior is that tied documents are returned in "index
order" but that can change as segments are merged
4) if you wish
: When I attempt the second example, of loading the *.xml files, I receive an
: error back. I tried just one of the XMLs and receive the same error.
Yeah ... there's a poor assumption here in the tutorial. note in
particular this paragraph...
--SNIP--
Solr's install includes a handful of So
: You can sort like this (I believe that _version_ is the internal id/index
: number for the document, but you might want to verify)
that is not true, and i strongly advise you not to try to sort on the
_version_ field ... for some queries/testing it may deceptively *look*
like it's sorting by
:
: Not sure I understand... _version_ is time based and hence will give
: roughly the same accuracy as something like
: TimestampUpdateProcessorFactory that you recommend below. Both
Hmmm... last time i looked, i thought _version_ numbers were allocated &
incremented on a per-shard basis and "
: I am not sure how to use "Sort By Function" for Case.
:
: |10#40|14#19|33#17|27#6|15#6|19#5|7#2|6#1|29#1|5#1|30#1|28#1|12#0|20#0|
:
: Can you tell how to fetch 40 when input is 10.
Something like...
if(termfreq(f,10),40,if(termfreq(f,14),19,if(termfreq(f,33),17,)))
But i suspect
Please note the exact description of hte property on the URL you
mentioned..
"The TZ parameter can be specified to override the default TimeZone (UTC)
used for the purposes of adding and rounding in date math"
The newer ref guide docs for this param also explain...
https://cwiki.apache.or
hat's a strainge error to get.
I can't explain why LinuxFileSystem can't load LinuxNativeDispatcher, but
you can probably bypass hte entire situation by explicitly configuring
ConcurrentMergeScheduler with defaults so that it doesn't try determine
wether you are using an SSD or "spinning" disk
: When I perform a range query of ['' TO *] to filter out docs where a
: particular field has a value, this does what I want, but I thought using the
: square brackets was inclusive, so empty-string values should actually be
: included?
I'm not sure i understand your question ... if you are deali
: > When I perform a range query of ['' TO *] to filter out docs where a
: > particular field has a value, this does what I want, but I thought using the
: > square brackets was inclusive, so empty-string values should actually be
: > included?
:
: They should be. Are you saying that zero length
: Can you do me a favour, I use solrJ to index, but I get all the
: Field is multivalued. How can I set my Field to not
: multivalued, can you tell me how to setting use solrJ.
If you are using a "Managed Schema" (which was explicitly configured in
most Solr 5.x exam
: I'm upgrading a plugin and use the AbstractSolrTestCase for tests. My tests
: work fine in 5.X but when I upgraded to 6.X the tests sometimes throw an
: error during initialization. Basically it says,
: "org.apache.solr.common.SolrException: Error instantiating
: shardHandlerFactory class
: [org
16 8:24 PM, "Chris Hostetter" wrote:
:
: >
: > : I'm upgrading a plugin and use the AbstractSolrTestCase for tests. My
: > tests
: > : work fine in 5.X but when I upgraded to 6.X the tests sometimes throw an
: > : error during initialization. Basically it says,
: > : &
the autoCommit settings initialize trackers so that they only fire after
some updates have been made -- don't think of it as a cron that fires
every X seconds, think of it as an update monitor that triggers timers.
if an update comes in, and there are no timers currently active, a timer
is cr
: a per field basis. I understand BM25 similarity is now supported in Solr
BM25 has been supported for a while, the major change recently is that it
is now the underlying default in Solr 6.
: but I was hoping to be able to configure k1 and b for different fields such
: as title, description, an
: Hi - bumping this issue. Any thoughts to share?
Shawn's response to your email seemed spot on acurate to me -- is there
something about his answer that doesn't match up with what you're seeing?
can you clarify/elaborate your concerns?
http://mail-archives.apache.org/mod_mbox/lucene-solr-user
I'm also confused by what exactly you mean by "doesn't work" but a general
suggestion you can try is putting the
RemoveBlankFieldUpdateProcessorFactory before your UUID Processor...
https://lucene.apache.org/solr/6_0_0/solr-core/org/apache/solr/update/processor/RemoveBlankFieldUpdateProcessorFa
The thing to remember about currency fields is that even if you tend to
only put one currency value in it, any question of interpreting the values
in that field has to be done relative to a specific currency, and the
exchange rates may change dynamically.
So use the currency function to get a
: At first, I saw the same exception you got ... but after a little while
: I figured out that this is because I was running the program more than
: once without deleting everything in the baseDir -- so the zookeeper
: server was starting with an existing database already containing the
: solr.xml
I can't explain the results you are seeing, but you also didn't provide us
with your schema.xml (ie; how are "text" and "text_auto" defined?) or
enough details ot try and reproduce on a small scale (ie: what does the
source data look like in the documents where these suggestion values
are comi
: I have a question for installing solr server. Using '
: install_solr_service.sh' with option -d , the solr home directory can be
: set. But the default log directory is under $SOLR_HOME/logs.
:
: Is it possible to specify the logs directory separately from solr home
directory during installa
: You could, but before that I'd try to see what's using your memory and see
: if you can decrease that. Maybe identify why you are running OOM now and
: not with your previous Solr version (assuming you weren't, and that you are
: running with the same JVM settings). A bigger heap usually means m
: Thanks, Nick. Do we know any suggested # for file descriptor limit with
: Solr6? Also wondering why i haven't seen this problem before with Solr 5.x?
are you running Solr6 on the exact same host OS that you were running
Solr5 on?
even if you are using the "same OS version" on a diff machine,
: > I’m getting this error on startup:
: >
: > section of solr.xml contains 1 unknown config parameter(s):
[shareSchema]
Pretty sure that's because it was never a supported property of the
section -- even in the old format of solr.xml.
it's just a top level property -- ie: create a child no
: HI All:I need a pagenigation with facet offset.
: There are two or more fields in [facet.pivot], but only one value
: for [facet.offset], eg: facet.offset=10&facet.pivot=field_1,field_2.
: In this condition, field_2 is 10's offset and then field_1 is 10's
: offset. Bu
: /?q=&wt=json&defType=dismax&q.alt=*:*&bq=provider:A^2.0/
: My first results have provider A.
: ?q=&wt=json&defType=dismax&q.alt=*:*&bq=provider:B^1.5
: My first results have provider B. Good!
: /?q=&wt=json&defType=dismax&q.alt=*:*&bq=provider:(A^2.0 B^1.5)/
: Then my first results h
Thanks to the SortedSetDocValues this is in fact possible -- in fact i
just uploaded a patch for SOLR-2522 that you can take a look at to get an
idea of how to make it work (the main class you're probably going
to want to look at is SortedSetSelector: you're going to want a similar
"SortedDocV
: Has anyone worked with deep pagination using SolrNet? The SolrNet
: version that I am using is v0.4.0.2002. I followed up with this article,
: https://github.com/mausch/SolrNet/blob/master/Documentation/CursorMark.md
: , however the version of SolrNet.dll does not expose the a StartOrCursor
:
:
:
: Most documents have a correctly formatted date string and I would like to keep
: that data available for search on the date field.
...
: I realize it is complaining because the date string isn't matching the
: data_driven_schema file. How can I coerce it into allowing the non-st
: I have a fresh install of Solr 5.2.1 with about 3 million docs freshly
: indexed (I can also reproduce this issue on 4.10.0). When I use the Solr
: MorelikeThisHandler with content stream I'm getting different results per
: shard.
I haven't looked at the code recently but i'm 99% certain that t
: > I am getting following exception for the query :
: > *q=field:query&stats=true&stats.field={!cardinality=1.0}field*. The
: > exception is not seen once the cardinality is set to 0.9 or less.
: > The field is *docValues enabled* and *indexed=false*. The same exception
: > I tried to reproduce o
https://people.apache.org/~hossman/#threadhijack
Thread Hijacking on Mailing Lists
When starting a new discussion on a mailing list, please do not reply to
an existing message, instead start a fresh email. Even if you change the
subject line of your email, other mail headers still track which
: My current expansion expands from the
:user-query
: to the
:+user-query favouring-query-depending-other-params overall-favoring-query
: (where the overall-favoring-query could be computed as a function).
: With the boost parameter, i'd do:
:(+user-query favouring-query-depending-othe
& configs you posted and see if you can reproduce using
the *exact* same data in two different fields, and if the choice of query
makes a differnce in the behavior you see.
:
: Regards,
: Modassar
:
: On Wed, Aug 19, 2015 at 10:24 AM, Modassar Ather
: wrote:
:
: > Ahmet/Chris! Thanks fo
: Can you please explain how having the same field for query and stat can
: cause some issue for my better understanding of this feature?
I don't know if it can, it probably shouldn't, but in terms of trying ot
udnerstand the bug and reproduce it, any pertinant facts may be relivant -
particula
: A follow up question. Is the sub-sorting on the lucene internal doc IDs
: ascending or descending order? That is, do the most recently index doc
you can not make any generic assumptions baout hte order of the internal
lucene doc IDS -- the secondary sort on the internal IDs is stable (and
F
1) The "terms" Query Parser (TermsQParser) has nothing to do with the
"TermsComponent" (the first is for quering many distinct terms, the
later is for requesting info about low level terms in your index)
https://cwiki.apache.org/confluence/display/solr/Other+Parsers#OtherParsers-TermsQueryParse
: updates? i can't do this because i have delta-import queries which also
: should be able to assign uuid when it is needed
You really need to give us a full and complete picture of what exactly you
are currently doing, what's working, what's not working, and when it's not
working what is it
: "Is there a way to search for documents that have a word appearing more
: than a certain number of times? For example, I want to find documents
: that only have more than 10 instances of the word "genetics" …"
Try...
q=text:genetics&fq={!frange+incl=false+l=10}termfreq('text','genetics')
No
Thats... strange.
Looking at hte code it appears to be a totally bogus and missleading
warning -- but it also shouldn't affect anything.
You can feel free to ignore it for now...
https://issues.apache.org/jira/browse/SOLR-7984
: Date: Thu, 27 Aug 2015 15:10:18 -0400
: From: Scott Hollenbeck
I just want to clarify: all of Shawn's points below are valid and good --
but they stll don't explain the warning messgae you are getting. it makes
no sense as the code is currently written, and doesn't do anything to help
encourage people to transition to path based handler names.
: Date:
: I have a "country" field in my index, with values like 'US', 'FR', 'UK',
: etc...
:
: Then I want our users to be able to define the order of their preferred
: countries so that grouped results are sorted according to their preference.
...
: Is there any other function that would allow
: various $HOME/solr-5.3.0 subdirectories. The documents/tutorials say to edit
: the solrconfig.xml file for various configuration details, but they never say
: which one of these dozen to edit. Moreover, I cannot determine which version
can you please give us a specific examples (ie: urls, page
: I'm trying to use localparams for stats component on Solr 4.4, exact query:
: q=*:*&core=hotel_reviews&collection=hotel_reviews&fq=checkout_date:[* TO
: *]&fq={!tag=period1}checkout_date:[2011-12-25T00:00:00.000Z TO
:
2012-01-02T00:00:00.000Z}&fq={!tag=period2}checkout_date:[2011-12-25T00:00:00
: Write a PostFilter which takes in a document id. It lets through all
: documents until it sees that document id. Once it sees it, it stops
: letting them through.
:
: Thus, the total count of documents would be the position of your queried
: car.
Sorry guys, that won't work.
PostFilter's can
depends on where you are reading "Process Total Time" from. that
terminology isn't something i've ever sen used in the context of solr
(fairly certain nothing in solr refers to anything that way)
QTime is the amount of time spent processing a request before it starts
being written out over th
: Strange enough, the following code gives different errors:
:
: assertQ(
I'm not sure what exactly assertQ will do in a distributed test like this
... probably nothing good. you'll almost certainly want to stick with
the distributed indexDoc() and query* methods and avoid assertU and
as
:
: how to parse json response from Solr Term Vector Component?
:
: I got following json structure from response when testing Solr 5.3.0
: tvComponent:
...
: Is it correct ? Why solr makes the json response for term vector
: information so difficult to extract from the client side ? why
You haven't really provided us enough info to make any meaningful
suggestions.
You've got at least 2 custom plugins -- but you don't give us any idea
what the implementations of those plugins look like, or how you've
configured them. Maybe there is a bug in your code? maybe it's
misconfigur
You're basically re-implementing Solr' cursors.
you can change your system of reading docs from the old collection to
use...
cursorMark=*&sort=timestamp+asc,id+asc
...and then instead of keeping track of the last timestamp & id values and
constructing a filter, you can just keep track of the
1 - 100 of 5498 matches
Mail list logo