I am trying to connect the SOLR with Java code using URLConnection, i have
deployed solr war file in jboss server(assuming server machine in some other
location or remote) its working fine if no exception raises... but if any
exception raises in server like connection failure its stopping the jboss
Am 26.10.2011 18:29, schrieb Shawn Heisey:
> For inserting, I do use a Collection of SolrInputDocuments. The delete
> process grabs values from idx_delete, does a query like the above (the
> part that's slow in Java), then if any documents are found, issues a
> deleteByQuery with the same string.
Sorry, I was wrong.
Am 27.10.2011 09:36, schrieb Michael Kuhlmann:
> and you'll get the number of affected documents in your response anyway.
That's not true, you don't get the affected document count. Anyway, it's
still true that you don't need to check for documents first, at least
not when you
> I am searching for 9065 , so its not
> about case sensitivity. My search is
> searching across all the field names and not limiting it to
> one
> field(specified in the qf param and using deftype dismax)
By saying case sensitivity, Erik was referring def*T*ype parameter itself. (not
the value o
--- On Wed, 10/26/11, darul wrote:
> From: darul
> Subject: Get results ordered by field content starting with specific word
> To: solr-user@lucene.apache.org
> Date: Wednesday, October 26, 2011, 11:36 PM
> I have seen many threads talking
> about it but not found any way on how to
> resolve i
Thanks Simon and Jay .That was helpful .
So what we are looking at during optimize is 2 or 3 times free Disk Space
to recreate the index.
Regards
Sujatha
On Wed, Oct 26, 2011 at 12:26 AM, Simon Willnauer <
simon.willna...@googlemail.com> wrote:
> RAM costs during optimize / merge is general
Well, at indexed time I can not touch because we do not have data to index
anymore.
To use SpanFirstQuery, I need to make a custom ParserQuery ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Get-results-ordered-by-field-content-starting-with-specific-word-tp3455754p3457167.
> Well, at indexed time I can not touch
> because we do not have data to index
> anymore.
>
> To use SpanFirstQuery, I need to make a custom ParserQuery
> ?
If re-index is not an option, then writing custom is necessary to use
SpanFirstQuery. You need to add it as an optional clause (with high b
hello,
I want to filter search by calendar availability. For each document I know
the days which it is not available.
How could I build my fields filter the documents that are available in a
range of dates?
For example, a document A is available from 1-9-2011 to 5-9-2011 and is
available from 17-9-
It appears that the solution to this is to ensure that the pattern for
your component field is longer than the pattern for your dynamic
"parent" field. This will ensure that the component field takes
precedence.
For example "*__coordinate" is longer than "OBJECT_LL_*" so it will take
precedence.
(11/10/27 9:12), Xue-Feng Yang wrote:
Hi,
From Solr Info page, I can see my solr-uima core is there, but
updateRequestProcessorChain is not there. What is the reason?
Because UpdateRequestProcessor(and Chain) is not type of SolrInfoMBean.
(As those classes in the page implement SolrInfoMBean
When we display search results to our users we include a percentage
score.
Top result being 100%, then all others normalised based on the
maxScore, calculated outside of Solr.
We now want to limit returned docs with a percentage score higher than
say, 50%.
e.g. We want to search but only r
what you is looking for is imho not releated to solr in special.
The topic should be solr as "temporal database".
In your case if you have a timeline from 0 to 10 and you have two
documents from 1 to 6 and 5 to 13 you can get all documents within 0 - 10
by quering document.end >= 0 and document.st
do your docs have daily availability ?
if so you could index each doc for each day (rather than have some
logic embedded in your data)
so instead of doc1 (1/9/2011 - 5/9/2011)
you have
doc1 1/9/2011
doc1 2/9/2011
doc1 3/9/2011
doc1 4/9/2011
doc1 5/9/2011
this makes search much easier and flexible
Have you tried varying mintf and mindf? Setting them higher than 1
seems like it would reduce the number of docs returned..
Best
Erick
On Tue, Oct 25, 2011 at 2:57 AM, vraa wrote:
> Hi
>
> I'm using the MoreLikeThis functionallity
> http://wiki.apache.org/solr/MoreLikeThis
> http://wiki.apa
From everything you've said, it certainly sounds like a low-level I/O
problem in the client, not a server slowdown of any sort. Maybe Perl is
using the same connection over and over (keep-alive) and Java is not. I
really don't know. One thing I've heard is that
StreamingUpdateSolrServer (I t
I have one query regarding solr search.I have one key words like "wireleess
mobilty kit" i need to search,I am not able to get when i am doing the
search.BUt when i have manually added in synonyms.txt file like[wirelss,
wireless access.etc] i am able to search the product related to this
.P
I've seen something like this done with an index of queries. That is, you
index actual user queries in some new core where each "document" is
a query. Then you issue the terms of the new query against this index
and get back similar "documents" (that are really queries). You'll want
to take some ca
Take a look at your admin/analysis page and put your tokens in for both
index and query times. What I think you'll see is that the # is being
stripped at query time due to the first PatternReplaceFilterFactory.
You probably want to split your analyzers into an index-time and query-time
pair and do
What happens if you change your WDDF definition in the query part of
your analysis
chain to NOT split on case change? Then your index should contain the right
fragments (and combined words) and your queries would match.
I admit I haven't thought this through entirely, but this would work
for your
What does your admin/analysis page show? And how about the
results with &debugQuery=on?
Best
Erick
On Wed, Oct 26, 2011 at 5:34 AM, roySolr wrote:
> Hi,
>
> I have some problems with the patternreplaceFilter. I can't use the
> worddelimiter because i only want to replace some special chars given
hi,
I have the following situation:
- A dropdownlist to search trips by Country
- A dropdownlist to search trips by departureperiod (range/month)
I want to have facetresults on these fields.
When i select a value in 1 of the dropdownlists, i receive the correct
numbers (facets)
If Country = "Belg
You've got two q parameters. For filtering on facet values, you're better off
using fq parameters instead (and if there is no other query, set q=*:*, or if
using dismax set q.alt=*:* and leave q empty/unspecified). Only one q
parameter is used, but any number of fq parameters may be specified.
Fantastic, thanks, yes I completely overlooked that case, separating the
analysers worked a treat.
Had also posted on stack overflow but the mailing list proved to be
superior!
Many thanks,
Daniel
On 27 October 2011 13:09, Erick Erickson wrote:
> Take a look at your admin/analysis page and pu
Hi Erik,
Thank you very much.
Your hint did solve the problem.
Acutally, i don't understand why (i read the difference between Q and QF,
but it's still not clear to me why it did'nt work with Q).
But it's solved, that's the most important :)
Thanks,
Ruben
--
View this message in context:
htt
I don't think it'll work as I've tried this approach myself and the blocking
issue was that Solr 1.4.1 use a different javabin version than Solr 3.4 (I
think it's 1 vs 2) so the master and the slave(s) can't communicate using
standard replication handler and thus can't exchange information and data
If we change the query chain to not split on case change, then we lose half the
benefit of that feature -- if a user types "WiFi" and the source record
contains "wi fi," we fail to get a hit. As you say, that may be worth
considering if it comes down to picking the lesser evil, but I still thin
Can you explain more what's the fieldType, what's the actual content of the
field in the document.
Why are you trying to use synonyms?
Regards
On Thu, Oct 27, 2011 at 7:55 AM, Sahoo, Jayanta wrote:
> I have one query regarding solr search.I have one key words like "wireleess
> mobilty kit" i nee
OK, so it sounds like the index.jsp welcome page setting is not the issue.
That is not a big surprise. (WebSphere does not have that as a global default,
but Jetty 6 certainly did, and it looks like Jetty 7 does as well).
BTW, that should be /solr/admin/index.jsp, as I indicated, not
/solr/a
Hi Robert,
take a look to
http://lucene.472066.n3.nabble.com/How-to-cut-off-hits-with-score-below-threshold-td3219064.html#a3219117
and
http://lucene.472066.n3.nabble.com/Filter-by-relevance-td1837486.html
So will
sort=date+desc&q={!frange l=0.85}query($qq)
qq=
help?
Best regards
Karsten
--
Sounds like a custom sorting collector would work - one that throws away docs
with less than some minimum score, so that it only collects/sorts documents
with some minimum score. AFAIK score is calculated even if you sort by some
other field.
On Oct 27, 2011, at 9:49 AM, karsten-s...@gmx.de wr
BTW, this would be good standard feature for SOLR, as I've run into this
requirement more than once.
On Oct 27, 2011, at 9:49 AM, karsten-s...@gmx.de wrote:
> Hi Robert,
>
> take a look to
> http://lucene.472066.n3.nabble.com/How-to-cut-off-hits-with-score-below-threshold-td3219064.html#a32191
I don't like the idea of indexing a doc per each value, the dataset can grow
a lot. I have thought that something like this could work:
At indexing time, if I know the dates of no avaliability, I could gather the
avaliability ones (will consider unknown as available). So, I index 4 fields
aval_yes_
On Thu, Oct 27, 2011 at 7:13 AM, Anatoli Matuskova <
anatoli.matusk...@gmail.com> wrote:
> I don't like the idea of indexing a doc per each value, the dataset can
> grow
> a lot.
What does a lot mean? How high is the sky?
A million people with 3 year schedules is a billion tiny documents.
Tha
> What does a lot mean? How high is the sky?
If I have 3 milion docs I would end up with 3 milion * days avaliable
> This can be done. And given that you want long stretches of availability,
> but what happens when a reservation is canceled? You have to coalesce
> intervals. That isn't impo
Hi guys,
I'm new to Solr (as you may guess for the subject). I'd like to force the
threshold for fuzzy queries to, say, 0.7. I've read that fuzzy queries are
expensive, but limiting it's threshold to a number near 1 would help.
So my question is: Is this possible to configure in some of the xml
c
On 10/27/2011 1:36 AM, Michael Kuhlmann wrote:
Why do you first query for these documents? Why don't you just delete
them? Solr won't harm if no documents are affected by your delete
query, and you'll get the number of affected documents in your
response anyway. When deleting, Solrj nearly does
I have a similar problem except I need to filter scores that are too high.
Robert Stewart 於 Oct 27, 2011 7:04 AM 寫道:
> BTW, this would be good standard feature for SOLR, as I've run into this
> requirement more than once.
>
>
> On Oct 27, 2011, at 9:49 AM, karsten-s...@gmx.de wrote:
>
>> H
Sorry my bad :(. Thanks for the help. It worked. I completely overlooked the
defType.
--
View this message in context:
http://lucene.472066.n3.nabble.com/DisMax-search-tp3455671p3458454.html
Sent from the Solr - User mailing list archive at Nabble.com.
I'm using the geohash field to store points for my data. When I do a
bounding box like:
localhost:8080/solr/select?q=point:[-45,-80%20TO%20-24,-39]
I get a data point that falls outside the box: (-73.03358 -50.46815)
The Spatial Search (http://wiki.apache.org/solr/SpatialSearch) pag
Hi guys,
If we ignore the features that Replication provides (
http://wiki.apache.org/solr/SolrReplication#Features),
which approach is better?
Is there any performance problems with Replication?
Replications seems quite easier (no special configuration, ssh setting, cron
setting),
while rsync is
I am not sure if there is such an option but you might be able to
override your query parser and reset that value if it is too fuzzy.
look for protected Query newFuzzyQuery(Term term, float
minimumSimilarity, int prefixLength) there you can change the actual
value used for minimumSimilarity
sim
On Thu, Oct 27, 2011 at 2:34 PM, Christopher Gross wrote:
> I'm using the geohash field to store points for my data. When I do a
> bounding box like:
>
> localhost:8080/solr/select?q=point:[-45,-80%20TO%20-24,-39]
>
> I get a data point that falls outside the box: (-73.03358
> -50.468155
True -- I found the geohash on a separate page. I was using it
because it can allow for multiple points, and I was hoping to be ahead
of the curve for allowing that feature for the data I'm managing.
I can roll back and use the LatLon type -- but then I'm still
concerned about the bounding box gi
Great! I didn't think there was a way to do it. I was about removing this
feature from my app for that reason. I'll give your advice it a try.
Thanks a lot!
2011/10/27 Simon Willnauer
> I am not sure if there is such an option but you might be able to
> override your query parser and reset tha
On Thu, Oct 27, 2011 at 3:22 PM, Christopher Gross wrote:
> I can roll back and use the LatLon type -- but then I'm still
> concerned about the bounding box giving results outside the specified
> range.
The implementation of things like bbox are intimately tied to the
field type (i.e. normally co
Meaning I need to implement my own QueryParser ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Get-results-ordered-by-field-content-starting-with-specific-word-tp3455754p3459064.html
Sent from the Solr - User mailing list archive at Nabble.com.
While sending documents with SolrJ Http API...at the end, I am never sure
documents are indexed.
I would like to store them somewhere and resend them in case commit has
failed.
If commit occurred every 10 minutes for example, and 100 documents are
waiting to be commit, server crash or stop..this
So Solr 1.4. I decided I wanted to change a field to have
omitNorms=true that didn't previously.
So I changed the schema to have omitNorms=true. And I reindexed all
documents.
But it seems to have had absolutely no effect. All relevancy rankings
seem to be the same.
Now, I could have a m
As far as I know there's no issue about this. You have to reindex and that's
it.
In which kind of field are you changing the norms? (You just will see
changes in text fields)
Using debugQuery=true you can see how norms affect the score (in case you
have them not omited)
--
View this message in con
Replication is easier to manage and a bit faster. See the performance
numbers: http://wiki.apache.org/solr/SolrReplication
--
View this message in context:
http://lucene.472066.n3.nabble.com/Collection-Distribution-vs-Replication-in-Solr-tp3458724p3459178.html
Sent from the Solr - User mailing li
I can't see those benchmarks, can you?
On Thu, Oct 27, 2011 at 5:20 PM, Marc Sturlese wrote:
> Replication is easier to manage and a bit faster. See the performance
> numbers: http://wiki.apache.org/solr/SolrReplication
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/C
I've been given the project of setting up a CentOS-based solr replication slave
for a project here at work. I think it's configured correctly, and replication
seems to be happening correctly.
I've got some CentOS experience, but I'm having to get up to speed on Solr in a
short period of time.
Two things:
1> Look at http://wiki.apache.org/solr/DataImportHandler, the
"interactive Development Mode" section. There's a page that helps you
debug this kind of thing. But I suspect your SQL is not correct. You
should be able to form a single SQL query that does what you want,
something like (and
Hmmm, I'm not sure what happens when you specify
(without type="index" and
. I have no clue which one
is used.
Look at the admin/analysis page to understand how things are
broken up.
Did you re-index after you added the ngram filter?
You'll get better help if you include example queries with
&d
we are not actively removing norms. if you set omitNorms=true and
index documents they won't have norms for this field. Yet, other
segment still have norms until they get merged with a segment that has
no norms for that field ie. omits norms. omitNorms is anti-viral so
once you set it to true it wi
On Thu, Oct 27, 2011 at 6:00 PM, Simon Willnauer
wrote:
> we are not actively removing norms. if you set omitNorms=true and
> index documents they won't have norms for this field. Yet, other
> segment still have norms until they get merged with a segment that has
> no norms for that field ie. omit
On 10/27/2011 5:56 AM, Michael Sokolov wrote:
From everything you've said, it certainly sounds like a low-level I/O
problem in the client, not a server slowdown of any sort. Maybe Perl
is using the same connection over and over (keep-alive) and Java is
not. I really don't know. One thing I'v
Does anyone have any idea on this issue?
On Tue, Oct 25, 2011 at 11:40 AM, Jason Toy wrote:
> Hi Yonik,
>
> Without a Join I would normally query user docs with:
> q=data_text:"test"&fq=is_active_boolean:true
>
> With joining users with posts, I get no no results:
> q={!join from=self_id_i
> to=
NP. By the way, kudos for posting enough information to diagnose
the problem first time round!
Erick
On Thu, Oct 27, 2011 at 8:46 AM, Daniel Bradley
wrote:
> Fantastic, thanks, yes I completely overlooked that case, separating the
> analysers worked a treat.
>
> Had also posted on stack overflow
Hmmm, this may be one of those things that's so ingrained it's not
mentioned. Certainly the CommonQueryParameters page never
explicitly says that there can only be one q parameter
But the problem is "how would multiple q params be combined"?
An implied AND? OR? NOT? the syntax would be a mess.
Thanks Koji,
I finally found a method not found error in SOLR 3.4. The method
resolveUpdateChainParam(SolrParams params, org.slf4j.Logger log) is not in the
class
org.apache.solr.util.SolrPluginUtils. It was very strange there were no errors
message. I found the problems after loaded source co
Would it be acceptable to change a central "slave config"? Because
it's possible to
have the replication process distribute solrconfig.xml files to the
slaves that are
different from the master.
That way, your master has it's own solrconfig.xml, and a solrconfig_slave.xml in
the conf directory. At
Thanks Koji,
I finally found a method not found error in SOLR 3.4. The method
resolveUpdateChainParam(SolrParams params, org.slf4j.Logger log) is not in the
class
org.apache.solr.util.SolrPluginUtils. It was very strange there were no errors
message. I found the problems after loaded source
Well, this could be explained if your fields are very short. Norms
are encoded into (part of?) a byte, so your ranking may be unaffected.
Try adding debugQuery=on and looking at the explanation. If you've
really omitted norms, I think you should see clauses like:
1.0 = fieldNorm(field=features, d
You've really got to give a lot more information about what you're trying to do
here, what you've tried and what you mean by "associate". Please review:
http://wiki.apache.org/solr/UsingMailingLists
Best
Erick
On Wed, Oct 26, 2011 at 6:29 PM, Milan Dobrota wrote:
> I created a custom plugin par
I am trying to highlight FieldA when a user searches on either FieldA or FieldB,
but I do not want to highlight FieldA when a user searches on FieldC.
To explain further: I have a field named "content" and a field named
"contentCS". The content field is a stored text field that uses
LowerCaseFilte
67 matches
Mail list logo