Henri,
look velocity.properties. I have there:
> input.encoding = UTF-8
Do you also?
This is the vm files' encodings.
Of course also make sure you edit these files in UTF-8 (using jEdit made it
trustable to me).
paul
Le 30 mars 2012 à 08:49, henri.gour...@laposte.net a écrit :
> OK
Hi all,
I am doing R&D about SOLR using any quantiles function for a set. I need a
quick-start road-map for modifying that quantiles function in my SOLR
plugin. I am thinking that it might be using any third party tools or
library for it.
--
Kashif Khan
--
View this message in context:
http://
OK, Ill try to provide more details:
I am using solr-3.5.0
I am running the example provided in the package.
Some of the modifications I have done in the various velocity/*.vm files
have accents!
It is those accents that show up garbled when I look at the results.
The .vm files are utf-8 encoded.
S
Why don't yu contribute RA to the source so that it is a
feature/module inside SOLR?
On Thu, Mar 29, 2012 at 8:32 AM, Nagendra Nagarajayya
wrote:
> It is from build 2012-03-19 from the trunk (part of the email). No fork.
>
>
> Regards,
>
> Nagendra Nagarajayya
> http://solr-ra.tgels.org
> http://
Can you also include a /select?q=*:*&wt=xml
?
On Thu, Mar 29, 2012 at 11:47 AM, Erick Erickson
wrote:
> Hmmm, looking at your schema, faceting on a really doesn't make
> all that much sense, there will always be exactly one of them. At
> least it's highly
> questionable.
>
> But that's not your
Ok.Then what does exactly zookeeper do in Solrcloud?Why we use?I am geetting
query response from both shards even without using zookeeper.
--
View this message in context:
http://lucene.472066.n3.nabble.com/SolrCloud-tp3867086p3869896.html
Sent from the Solr - User mailing list archive at Nabble.
I don't believe this is related to that bug as I don't set the facet
prefix. I have honestly never seen the issue before either and an
optimize fixed it. I can try to see if I can duplicate the issue
tomorrow, but since I've done the optimize I haven't seen it again.
On Thu, Mar 29, 2012 at 7:45
I am using the function query as part of an frange.
So, something like this: q=productId:[* TO *] fq={!frange
l=1}ceil(query(!v='documentType:(blah1 blah2 blah3)'))
This is actually quite slow. I suspect that the problem is the query
function is calculating a score for every document in the inde
On Thu, Mar 29, 2012 at 6:33 PM, Jamie Johnson wrote:
> I recently got this stack trace when trying to execute a facet based
> query on my index. The error went away when I did an optimize but I
> was surprised to see it at all. Can anyone shed some light on why
> this may have happened?
I don'
The documents are removed from the search when the delete is committed.
The space for those documents is reclaimed at the next merge for the segment
where they were.
wunder
On Mar 29, 2012, at 4:15 PM, Jamie Johnson wrote:
> Thanks, does it matter that we are also updates to documents at
> va
On Thu, Mar 29, 2012 at 7:15 PM, Jamie Johnson wrote:
> Thanks, does it matter that we are also updates to documents at
> various times? Do the deleted documents get removed when doing a
> merge or does that only get done on an optimize?
Yes, any merge removes documents that have been marked as
Thanks, does it matter that we are also updates to documents at
various times? Do the deleted documents get removed when doing a
merge or does that only get done on an optimize?
On Thu, Mar 29, 2012 at 7:08 PM, Walter Underwood wrote:
> Don't. "Optimize" is a poorly-chosen name for a full merge.
Don't. "Optimize" is a poorly-chosen name for a full merge. It doesn't make
that much difference and there is almost never a need to do it on a periodic
basis.
The full merge will mean a longer time between the commit and the time that the
data is first searchable. Do the commit, then search.
On Thu, Mar 29, 2012 at 6:44 PM, Alexandre Rocco wrote:
> Yonik,
>
> Thanks for the heads-up. That one worked.
>
> Just trying to wrap around how it would work on a real case. To test this
> one I just got the coordinates from Google Maps and searched within the pair
> of coordinates as I got them
Yonik,
Thanks for the heads-up. That one worked.
Just trying to wrap around how it would work on a real case. To test this
one I just got the coordinates from Google Maps and searched within the
pair of coordinates as I got them. Should I always check which is the lower
and upper to assemble the
On Thu, Mar 29, 2012 at 6:20 PM, Alexandre Rocco wrote:
> http://localhost:8984/solr/select?q=*:*&fq=local:[-23.6677,-46.7315 TO
> -23.6709,-46.7261]
Range queries always need to be [lower_bound TO upper_bound]
Try
http://localhost:8984/solr/select?q=*:*&fq=local:[-23.6709,-46.7315 TO
-23.6677,-4
Erick,
Just checked on the separate fields and everything looks fine.
One thing that I'm not completely sure is if this query I tried to perform
is correct.
One sample document looks like this:
200
-23.6696784,-46.7290193
-23.6696784
-46.7290193
So, to find for this document I tried to create
Yeah, I guess that would work. I wasn't sure if it would change relative
to other documents. But if it were to be combined with other fields,
that approach may not work because the calculation wouldn't include the
scoring for other parts of the query. So then you have the dynamic score
and what to
I don't think there's really any reason SolrCloud won't work with
Tomcat, the setup is
probably just tricky. See:
http://lucene.472066.n3.nabble.com/SolrCloud-new-td1528872.html
It's about a year old, but might prove helpful.
Best
Erick
On Thu, Mar 29, 2012 at 3:41 PM, Vadim Kisselmann
wrote:
>
Oops... my previous replies accidentally went off-list. I'll cut-n-paste below.
OK, so it looks like there is probably no bug here - it's simply that
commits can sometimes take a long time and updates were blocked during
that time (and would have succeeded eventually except the jetty
timeout was
Yes, i think so, too :)
MLT doesn´t need termVectors really, but it´s faster with them. I
found out, what
MLT works better on the title field in my case, instead of big text fields.
Sharding is in planning, but my setup with SolrCloud, ZK and Tomcat
doesn´t work,
see here:
http://mail-archives.ap
Can't you simply calculate that at index time and assign the result to a
field, then sort by that field.
On Thu, Mar 29, 2012 at 12:07 PM, Darren Govoni wrote:
> I'm going to try index time per-field boosting and do the boost
> computation at index time and see if that helps.
>
> On Thu, 2012-03
I'm currently looking to see what would be a decent way to implement a
scrolling window in the result set when looking for an item.
Basically, I need to find item X in the result set and return say N items
before and N items after.
< - N items -- Item X --- N items >
So I was thinking a
This all looks fine, so the next question is whether or not your
documents have the value you think.
+local_0_coordinate:[-23.6674 TO -23.6705] +local_1_coordinate:[-46.7314 TO
-46.7274]
is the actual translated filter.
So I'd check the actual documents in the index to see if you have a single
do
Boosting won't help either I don't think. Boosts apply to the
document, and pretty much ignore position information.
Best
Erick
On Thu, Mar 29, 2012 at 2:07 PM, Abhishek tiwari
wrote:
> can i achieve this with help of boosting technique ?
>
> On Thu, Mar 29, 2012 at 10:42 PM, Erick Erickson
>
Erick,
My location field is defined like in the example project:
Also, there is the dynamic that stores the splitted coordinates:
The response XML with debugQuery=on is looking like this:
0
1
*:*
*:*
MatchAllDocsQuery(*:*)
*:*
LuceneQParser
local:[-23.6674,-46.7314 TO -23.6705,-46.7274]
Yonik, I didn't say there was an update request active at the moment the
thread dump was made, only that previous update requests failed with a
timeout. So maybe this is the missing piece.
I didn't enable nio with Jetty, probably it's there by default. Disabling
it is the next thing to check.
If so
On Thu, Mar 29, 2012 at 1:50 PM, Rafal Gwizdala
wrote:
> Below i'm pasting the thread dump taken when the update was hung (it's also
> attached to the first message of this topic)
Interesting...
It looks like there's only one thread in solr code (the one generating
the thread dump).
The stack tr
can i achieve this with help of boosting technique ?
On Thu, Mar 29, 2012 at 10:42 PM, Erick Erickson wrote:
> Solr doesn't support sorting on multValued fields so I don't think this
> is possible OOB.
>
> I can't come up with a clever indexing solution that does this either,
> sorry.
>
> Best
>
More memory is not necessarily better, it can lead to longer, more
intense garbage collections that cause things to stop. You might
also consider lowering your memory allocation, but 2G is really not
all that much so I somewhat doubt it's a problem but thought I'd
mention it.
Best
Erick
On Thu, M
What are your results? Can you show us the field definition for "local"
and the results of adding &debugQuery=on?
Because this should work as far as I can tell.
Best
Erick
On Thu, Mar 29, 2012 at 11:04 AM, Alexandre Rocco wrote:
> Hello,
>
> I'm trying to perform some queries on a location fiel
Guys, thanks for all the suggestions
I will be trying them, one at a time. Imho it's too early to give up and
look for another tool, I'll try to work on configuration and see what
happens.
The NRT looks quite promising, there are also tons of config options to
change.
As for now, I have made the up
Hmmm, looking at your schema, faceting on a really doesn't make
all that much sense, there will always be exactly one of them. At
least it's highly
questionable.
But that's not your problem and what's wrong isn't at all obvious. Can you try
pasting the results of adding &debugQuery=on?
Best
Eric
I doubt that the pre-installed Jetty server has problems with UTF-8, although
you haven't told us what version of Solr you're running on so it could be really
old.
And you also haven't told us why you think UTF-8 is a problem. How is this
manifesting itself? Failed searches? Failed indexing? ???
: 6403279 Nov 22 14:54 apache-solr-3.5.0.war
:
: when i build the war file from source - i get a different sized file:
:
: ./dist/apache-solr-3.5-SNAPSHOT.war
:
: 6404098 Mar 29 11:41 ./dist/apache-solr-3.5-SNAPSHOT.war
:
: am i building from the wrong source?
I haven't tested this to confir
Yeah, it's worth a try. The term vectors aren't entirely necessary for
highlighting,
although they do make things more efficient.
As far as MLT, does MLT really need such a big field?
But you may be on your way to sharding your index if you remove this info
and testing shows problems
Best
Er
Solr doesn't support sorting on multValued fields so I don't think this
is possible OOB.
I can't come up with a clever indexing solution that does this either, sorry.
Best
Erick
On Thu, Mar 29, 2012 at 8:27 AM, Abhishek tiwari
wrote:
> a) No. i do not want to sort the content within document .
hello all,
i have been pulling down the 3.5 solr war file from the mirror site.
the size of this file is:
6403279 Nov 22 14:54 apache-solr-3.5.0.war
when i build the war file from source - i get a different sized file:
./dist/apache-solr-3.5-SNAPSHOT.war
6404098 Mar 29 11:41 ./dist/apache-s
Thanks for the tips, but unfortunately, no progress so far.
Reading through the Web, I guess that jetty has utf-8 problems!
I guess that I will have to switch from the embedded (and pre installed ->
easy) jetty server present in Solr in favor of Tomcat (for which I have to
rediscover the installati
Hello
I am trying to filter out characters per unicode block or before
tokenization, so I use "PatternReplaceCharFilterFactory". In the end, I want
to filter out all non-CJK characters, basically latin, greek, arabic and
hebrew scripts.
The problem is, PatternReplaceCharFilterFactory does not ful
The version is 3.5.0.2011.11.22.14.54.38. I did not apply any patches, but
then again it is not my server.
Do you have a clue on what is going wrong here?
Regards,
Youri
2012/3/29 Bill Bell
> Send schema.xml and did you apply any patches? What version of Solr?
>
> Bill Bell
> Sent from mobile
I'm going to try index time per-field boosting and do the boost
computation at index time and see if that helps.
On Thu, 2012-03-29 at 10:08 -0400, Darren Govoni wrote:
> Hi,
> I have a situation I want to re-score document relevance.
>
> Let's say I have two fields:
>
> text: The quick brown f
Hello,
I'm trying to perform some queries on a location field on the index.
The requirement is to search listings inside a pair of coordinates, like a
bounding box.
Taking a look on the wiki, I noticed that there is the option to use the
bbox query but in does not create a retangular shaped box t
Also, in case you use Apache's mod_proxy, be sure to use the nocanon attribute.
(I don't know of an equivalent for mod_rewrite).
In general, I tend also to advise also to change the default encoding of the
java running the servlets... but I am sure you've done this.
Tell us your success or lack
Hi, Henri.
Make sure that the container in which you are running Solr is also set for
UTF-8.
For example, in Tomcat, in the server.xml file, your Connector definitions
should include:
URIEncoding="UTF-8"
Bob Sandiford | Lead Software Engineer | SirsiDynix
P: 800.288.8020 X6943 | bob.sa
I was wrong! It does seem to work!
Thanks a bunch!
cheers,
:-Dennis
On Mar 29, 2012, at 15:52 , fbrisbart wrote:
> I had the same issue months ago.
> 'newSearcher' fixed the problem for me.
> I also remember that I had to upgrade solr (3.1) because it didn't work
> with release 1.4
> But, I
I cant get utf-8 encoding to work!!
I havetext/html;charset=UTF-8
in my request handler, and
input.encoding=UTF-8
output.encoding=UTF-8
in velocity.properties, in various locations (I may have the wrong ones! at
least in the folder where the .vm files reside)
What else should I be doing
If you must have real-time search, you might look at systems that are designed
to do that. MarkLogic isn't free, but it is fast and real-time. You can use
their no-charge Express license for development and prototyping:
http://developer.marklogic.com/express
OK, back to Solr.
wunder
Search Guy
It is from build 2012-03-19 from the trunk (part of the email). No fork.
Regards,
Nagendra Nagarajayya
http://solr-ra.tgels.org
http://rankingalgorithm.tgels.org
On 3/29/2012 7:20 AM, Bernd Fehling wrote:
Nothing against RankingAlgorithm and your work, which sounds great, but
I think that YOUR
You can also use $deleteDocById . If you also use $skipDoc, you can sometimes
get the deletes on the same entity with a "command=full-import&clean=false"
delta. This may or may not be more convienent that what you're doing already.
See http://wiki.apache.org/solr/DataImportHandler#Special_Comm
Nothing against RankingAlgorithm and your work, which sounds great, but
I think that YOUR "Solr 4.0" might confuse some Solr users and/or newbees.
As far as I know the next official release will be 3.6.
So your "Solr 4.0" is a trunk snapshot or what?
If so, which revision number?
Or have you do
On Thu, Mar 29, 2012 at 4:24 AM, Lance Norskog wrote:
> 5-7 seconds- there's the problem. If you want to have documents
> visible for search within that time, you want to use the trunk and
> "near-real-time" search. A hard commit does several hard writes to the
> disk (with the fsync() system call
Send schema.xml and did you apply any patches? What version of Solr?
Bill Bell
Sent from mobile
On Mar 29, 2012, at 5:26 AM, Youri Westerman wrote:
> Hi,
>
> I'm currently learning how to use solr and everything seems pretty straight
> forward. For some reason when I use faceted queries it re
Hi,
I have a situation I want to re-score document relevance.
Let's say I have two fields:
text: The quick brown fox jumped over the white fence.
terms: fox fence
Now my queries come in as:
terms:[* TO *]
and Solr scores them on that field.
What I want is to rank them according to the distr
Have you tried using Solr 3.5 with RankingAlgorithm 1.4.1 ? Has NRT
support and is very fast, updates about 5000 documents in about 490 ms
(while updating 1m docs in batches of 5k).
You can get more info from here:
http://solr-ra.tgels.com/wiki/en/Near_Real_Time_Search_ver_3.x
Regards,
Nagen
I had the same issue months ago.
'newSearcher' fixed the problem for me.
I also remember that I had to upgrade solr (3.1) because it didn't work
with release 1.4
But, I suppose you already have a solr 3.x or more.
So I'm afraid I can't help you more :o(
Franck
Le jeudi 29 mars 2012 à 15:41 +020
I am very excited to announce the availability of Solr 4.0 with
RankingAlgorithm 1.4.1 (NRT support) (build 2012-03-19). The NRT
implementation now supports both RankingAlgorithm and Lucene.
RankingAlgorithm 1.4.1 has improved performance over the earlier release
(1.4) and supports the entire
On Mar 29, 2012, at 14:49 , fbrisbart wrote:
> Arf, I didn't see your attached tgz.
>
> In your slave solrconfig.xml, only the 'firstSearcher' contains the
> query. Add it also in the 'newSearcher', so that the new search
> instances will wait also after a new index is replicated.
Did that now,
Hi Erick,
thanks:)
The admin UI give me the counts, so i can identify fields with big
bulks of unique terms.
I known this wiki-page, but i read it one more time.
List of my file extensions with size in GB(Index size ~150GB):
tvf 90GB
fdt 30GB
tim 18GB
prx 15GB
frq 12GB
tip 200MB
tvx 150MB
tvf is m
On 3/29/2012 2:49 AM, Rafal Gwizdala wrote:
That's bad news.
If 5-7 seconds is not safe then what is the safe interval for updates?
Near real-time is not for me as it works only when querying by document Id
- this doesn't solve anything in my case. I just want the index to be
updated in real-time
We're doing something similar: We want to combine search relevancy with a
fitness value computed from several other data sources.
For this, we pre-compute the fitness value for each document and store it a
flat file (lines of the format document_id=fitness_score) that we use an
externalFileField
Arf, I didn't see your attached tgz.
In your slave solrconfig.xml, only the 'firstSearcher' contains the
query. Add it also in the 'newSearcher', so that the new search
instances will wait also after a new index is replicated.
The first request is long because the default faceting method uses the
Yes I figured my problem would be something like that. I'll try with the
catalina/tomcat home variables I think.
Thank you.
-Original Message-
From: Ingar Hov [mailto:ingar@gmail.com]
Sent: 29. mars 2012 14:13
To: solr-user@lucene.apache.org
Subject: Re: Build solr with Maven
I think
I do have a firstSearcher, but currently coldSearcher is set to true. But
doesn't this just mean that that any searches will block while the first
searcher is running? This is how the comment describes first searcher. It would
almost give the same effect; that some searches take a long time.
Hi,
I am trying to do the gzip compression in solr server.
I had refered to the below link to add the same in the web.xml
http://blog.max.berger.name/2010/01/jetty-7-gzip-filter.html
I am using jetty server version 6. When I restart the server after adding
the above changes, I got the followi
Could be garbage collection. Could be larger and larger merges. At some point
your commit will cause all segments to be merged. It's likely that what's
happening is you need to hit the "magic combination" of events, particularly
the problem of too many warming searchers.
So, look at your log files
a) No. i do not want to sort the content within document .
I want to sort the documents .
b) As i have explained i have result set( documents ) and each document
contains a fields "*ad_text*" (with other fields also) which is
multivalued..storing some tags say "B1, B2, B3" in each. bt order of tags
This is the way SolrCloud works at present. There must be at least
one instance of each shard up in order to get results. I believe there
are plans to return partial results in future, but that's not been
implemented yet.
Best
Erick
On Thu, Mar 29, 2012 at 4:37 AM, asia wrote:
> Hello,
> I am wo
I guess my question is "why are you using a catchall field at all"? This
is the kind of thing edismax was designed for, so your qf could just
contain all the fields with appropriate boosts, there aren't that many...
But what you're actually doing will probably work. I think if you're not
seeing DE
I think you need absolute path. But perhaps if $CATALINA_HOME or
$TOMCAT_HOME is set, you can use it with your path.Haven't tried it
though..
In any case, you should quite easily be able to verify if relative
paths can be used. All you need to do is to get the work directory for
the webapp and ref
One option to try here (not verified) is to set up a Solr front that will
point to these two shards. Then try accessing its luke interface via admin
as you did on one of the shards.
But as Erick already pointed out, Luke operates on a lower level than Solr,
so this does not necessarily work.
Dmit
Alexander:
Your images were stripped by one of our mail servers, so there's not
much we can see ...
But guessing, you aren't searching the fields you think you are:
itemNameSearch:fifa12
becomes
itemNameSearch:fifa defaultSearchField:12
where defaultSearchField is defined in your schema.xml file
If you add your query to the firstSearcher and/or newSearcher event
listeners in the slave
'solrconfig.xml' (
http://wiki.apache.org/solr/SolrCaching#newSearcher_and_firstSearcher_Event_Listeners
),
each new search instance will wait before accepting queries.
Example to load the FieldCache for
I very much doubt that you can persuade Luke to reach across shards. Shards
are really a higher-level notion, the automatic distribution of requests
across shards is really a Solr-level construct (making use of the lower-
level Lucene capabilities, to be sure). With Luke, you point
it at index file
The admin UI (schema browser) will give you the counts of unique terms
in your fields, which is where I'd start.
I suspect you've already seen this page, but if not:
http://lucene.apache.org/java/3_5_0/fileformats.html#file-names
the .fdt and .fdx file extensions are where data goes when
you set '
Hmmm, I don't quite get this. Are you saying that you want
to sort the documents or sort the content within the document?
Sorting documents (i.e the results list) requires a single-valued
field. So you'd have to, at index time, sort the entries.
Sorting the content within the document is somethin
Tried that, but I guess I am doing it wrong somehow with the paths.
The home folder should be WEB-INF/solr inside the tomcat. But how would I
set that path correctly? Do I need to use absolute paths?
-Original Message-
From: Ingar Hov [mailto:ingar@gmail.com]
Sent: 29. mars 2012 12:5
Hi,
I'm currently learning how to use solr and everything seems pretty straight
forward. For some reason when I use faceted queries it returns only empty
sets in the facet_count section.
The get params I'm using are:
?q=*:*&rows=0&facet=true&facet.field=urn
The result:
"facet_counts": {
Hello everyone!
I am new to Solr and I have been doing a bit of reading about boosting
search results. My search index consists of products with different
attributes like a title, a description, a brand, a price, a discount percent
and so on. I would like to do a fairly complex boosting, so that f
- Forwarded Message -
From: Rahul Mandaliya
To: "solr-user@lucene.apache.org"
Sent: Thursday, March 29, 2012 9:38 AM
Subject: Fw: confirm subscribe to solr-user@lucene.apache.org
hi,
i am giving confirmation for subscribtion to solr-user@lucene.apache.org
regards,
Rahul
__
I see..
Try to use ... in web.xml.
Regards,
Ingar
On Thu, Mar 29, 2012 at 8:34 AM, Aleksander Akerø
wrote:
> Well, it's got all to do with how we have decided the rest of our deployment
> environment. So the point is basicly that there should be no configurations
> to the tomcat because the web
@All
Why nobody desires such a pretty cool feature?
Nicholas,
I have a tiny progress: I'm able to stream in javabin codec format while
searching, It implies sorting by _docid_
here is the diff
https://github.com/m-khl/solr-patches/commit/2f9ff068c379b3008bb983d0df69dff714ddde95
The current issue
That's bad news.
If 5-7 seconds is not safe then what is the safe interval for updates?
Near real-time is not for me as it works only when querying by document Id
- this doesn't solve anything in my case. I just want the index to be
updated in real-time, 30-40 seconds delay is acceptable but not mu
Hello,
I am working on solr.I have set up 2 solr instances on different systems i.e
i did sharding.I am using tomcat and eclipse environment.When I fire a query
in solrj for a data from index,i get response when both system's tomcat is
working.But when I stop one of the system's server I dont get
5-7 seconds- there's the problem. If you want to have documents
visible for search within that time, you want to use the trunk and
"near-real-time" search. A hard commit does several hard writes to the
disk (with the fsync() system call). It does not run smoothly at that
rate. It is no surprise tha
On 29 March 2012 13:35, elisabeth benoit wrote:
> Hi all,
>
> I'm using solr 3.4 with a catchall field and an edismaw request handler.
> I'd like to score higher answers matching with words not contained in one
> of the fields copied into my catchall field.
>
> So my catchallfield is called catcha
Hi all,
I'm using solr 3.4 with a catchall field and an edismaw request handler.
I'd like to score higher answers matching with words not contained in one
of the fields copied into my catchall field.
So my catchallfield is called catchall. It contains, let's say, fields
NAME, CATEGORY, TOWN, WAY
Hi,
You should disable velocity by adding -Dsolr.velocity.enabled=false to
JAVA_OPTS
--Jamel
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Tomcat-Install-tp3865290p3866947.html
Sent from the Solr - User mailing list archive at Nabble.com.
88 matches
Mail list logo