but if I use this field to do sorting, there will be an error occured
and throw an indexOfBoundArray exception.
On Thursday, November 11, 2010, Robert Petersen wrote:
> 1) Just put the new field in the schema and stop/start solr. Documents
> in the index will not have the field until you reinde
On Nov 10, 2010, at 12:30pm, Bob Sandiford wrote:
Why not use replication? Call it inexperience...
We're really early into working with and fully understanding Solr
and the best way to approach various issues. I did mention that
this was a prototype and non-production code, so I'm covere
On Wed, Nov 10, 2010 at 8:29 PM, Chris Hostetter
wrote:
>
> (forwarded on behalf of robo ... trying to figured out odd spam blocking
> issue he's having)
>
> -- Forwarded message --
>
> Good day,
>
> We are new to Solr and trying to setup a HA configuration in the
> cloud. We hav
(forwarded on behalf of robo ... trying to figured out odd spam blocking
issue he's having)
-- Forwarded message --
Good day,
We are new to Solr and trying to setup a HA configuration in the
cloud. We have setup a Solr Master server which does all the
indexing. We have 2 So
On Nov 11, 2010, at 1:12 AM, Jonathan Rochkind wrote:
> Are you sure you really want to throw out stopwords for your use case? I
> don't think autocompletion will work how you want if you do.
in our case i think it makes sense. the content is targetting the electronic
music / dj scene, so we
Are you sure you really want to throw out stopwords for your use case? I don't
think autocompletion will work how you want if you do.
And if you don't... then why use the WhitespaceTokenizer and then try to jam
the tokens back together? Why not just NOT tokenize in the first place. Use the
Ke
Hi,
i've created the following filterchain in a field type, the idea is to use it
for autocompletion purposes:
With that kind of filterchain, the EdgeNGramFilterFactory will receive multiple
tokens on input strings with whitespaces in it. This leads to the following
results:
I
Thank you for your answers.
Isn't it possible to tune Solr to use less disc bandwidth (involving a
longer optimization) ?
I moved Solr on the unused HDD, and the problem is solved !
Fortunately I have this separate disk...
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr
Good call. Alternately, for facet limiting, you may find it simpler,
easier (and very very slightly more efficient for Solr) to use either
the "raw" or "field" query parsers, that don't do the "pre-tokenization"
that the standard query parser does, which is what is making the quotes
required.
O wow, the quotes did the trick...thanks! :)
--
View this message in context:
http://lucene.472066.n3.nabble.com/Facet-showing-MORE-results-than-expected-when-its-selected-tp1878828p1879335.html
Sent from the Solr - User mailing list archive at Nabble.com.
I was playing around with the example Solr app, and I get different
results when I specify something like fq=manu_exact:"ASUS Computer Inc."
and fq=manu_exact:ASUS Computer Inc.
The latter gives many more matches, which looks kinda familiar.
Silly season stuff, I know, but thought I'd mention
Nope, thats not possible either since the themename is stored in the database
in table [themes] only once, and other locations refer to it using the link
table [location_themes]
simple DB scheme using a link table:
[themes]
id name
[location_themes]
locationid themeid
[locations]
id name etc e
Another option : assuming themes_raw is type 'string' (couldn't get that
nugget of info for 100%) it could be that you're seeing a difference in nr
of results between the 110 for fq:themes_raw and 321 from your db, because
fieldtype:string (thus themes_raw) is case-sensitive while (depending on
y
Nope, I restarted my server to reload schema.xml, and did a reindex, as I've
done a thousand times before, but still the same behaviour :(
--
View this message in context:
http://lucene.472066.n3.nabble.com/Facet-showing-MORE-results-than-expected-when-its-selected-tp1878828p1879218.html
Sent fr
I've had that sort of thing happen from 'corrupting' my index, by
changing my schema.xml without re-indexing.
If you change field types or other things in schema.xml, you need to
reindex all your data. (You can add brand new fields or types without
having to re-index, but most other changes wi
LOL, very clever indeed ;)
The thing is: when I select the amount of records matching the theme 'Hotel
en Restaurant' in my db, I end up with 321 records. So that is correct. I
dont know where the 370 is coming from.
Now when I change the query to this: &fq=themes_raw:Hotel en Restaurant
I end
Hi
I'm fairly new to solr but I have it configured, along with nutch, as per
this tutorial http://ubuntuforums.org/showthread.php?p=9596257.
Nutch is crawling and injecting documents into solr as expected, however, I
want to break the data down further so what ends up in solr is a bit more
granul
Shouldn't the second query have the clause:
&fq=themes_raw:Hotel en Restaurant
instead of:
&fq=themes:Hotel en Restaurant
Otherwise you're mixing apples (themes_raw) and oranges (themes).
(Notice how I cleverly extended the restaurant theme to be food related :))
Bob Sandiford | Lead So
A facet shows the amount of results that match with that facet, e.g. "New
York (433)" So when the facet is clicked, you'd expect that amount of
results (433).
However, I have a facet "Hotel en Restaurant (321)", that, when clicked
shows 370 results! :s
1st query:
http://localhost:8983/solr/db/s
1) Just put the new field in the schema and stop/start solr. Documents
in the index will not have the field until you reindex them but it won't
hurt anything.
2) Just turn off their handlers in solrconfig is all I think that
takes.
-Original Message-
From: gauravshetti [mailto:gaurav.s
thx so much tomas, I'll test now.
On Wed, Nov 10, 2010 at 6:47 PM, Tomas Fernandez Lobbe <
tomasflo...@yahoo.com.ar> wrote:
> You have to modify the field type you are using in your schema.xml file.
> This is
> the "text" field type of Solr 1.4.1 exmple with this filter added:
>
> positionI
In the header is a line saying what rules your message matched. That'll
let you know what about your message was causing your mails to be
rejected.
Upayavira
On Wed, 10 Nov 2010 11:42 -0800, "robo -" wrote:
> Thanks for all your help Ezequiel. I cannot see anything in my email
> that would make
You have to modify the field type you are using in your schema.xml file. This
is
the "text" field type of Solr 1.4.1 exmple with this filter added:
Ok tks,
I'm new with solr, my doubt is how can I enable theses feature. Or these
feature is already working by default?
Is this something to config on my schema.xml?
Tks!!
On Wed, Nov 10, 2010 at 6:40 PM, Tomas Fernandez Lobbe <
tomasflo...@yahoo.com.ar> wrote:
> That's what the ASCIIFoldingF
That's what the ASCIIFoldingFilter does, it removes the accents, that's why you
have to add it to the query analisis chain and to the index analysis chain, to
search the same way you index.
You can see how it works from the Analysis page on Solr Admin.
De
Hi Tomas,
Do you have some example to put in schema.xml?
How can I use thes filter class?
Tks
On Wed, Nov 10, 2010 at 6:25 PM, Tomas Fernandez Lobbe <
tomasflo...@yahoo.com.ar> wrote:
> It looks like ISOLatin1AccentFilter is deprecated on Solr 1.4.1, If you are
> on
> that version, you should
Why not use replication? Call it inexperience...
We're really early into working with and fully understanding Solr and the best
way to approach various issues. I did mention that this was a prototype and
non-production code, so I'm covered, though :)
We'll take a look at the replication featu
have you tried using a TokenFilter which removes accents both at
indexing and searching time? If you index terms without accents and
search the same
way you should be able to find all documents as you require.
On 10 November 2010 20:25, Tomas Fernandez Lobbe
wrote:
> It looks like ISOLatin1Acce
You could use the actual built-in Solr replication feature to accomplish
that same function -- complete re-index to a 'master', and then when
finished, trigger replication to the 'slave', with the 'slave' being the
live index that actually serves your applications.
I am curious if there was an
It looks like ISOLatin1AccentFilter is deprecated on Solr 1.4.1, If you are on
that version, you should use the ASCIIFoldingFilter instead.
Like with any other filter, to use it, you have to add the filter factory to
the
analysis chain of the field type you are using:
Make sure you add it to
have you tried using a TokenFilter which removes accents both at indexing
and searching time? If you index terms without accents and search the same
way you should be able to find all documents as you require.
On 10 November 2010 20:08, Claudio Devecchi wrote:
> Tomas,
>
> Let me try to explain
: I built the trunk and deploy the war, but cannot access the admin URL
: anymore.
:
: Error loading class
: 'org.apache.solr.highlight.MultiColoredScoreOrderFragmentsBuilder
:
: This class seems to be missing?
You appear to be using an "old" copy of the example config that refrences
a class
Tomas,
Let me try to explain better.
For example.
- I have 10 documents, where 7 have the word pereque (without accent) and 3
have the word perequê (with accent)
When I do a search pereque, solr is returning just 7, and when I do a search
perequê solr is returning 3.
But for me, these words ar
We also use SolrJ, and have a dynamically created Core capability - where we
don't know in advance what the Cores will be that we require.
We almost always do a complete index build, and if there's a previous instance
of that index, it needs to be available during a complete index build, so we
: As far as I know, in the core admin page you can find when was the last time
: an index had a modification and was comitted checking the lastModified.
: But? what startTime and uptime mean?
: Thanks in advance
startTime should be when the core was created (ie: when it started) uptime
is now-st
: There is now a data/index with a write lock file in it. I have not
: attempted to read the index, let alone add something to it.
: I start solr again, and it cannot open the index because of the write lock.
Lance, i can't reproduce using trunk r1033664 on Linux w/ext4 -- what OS &
Filesystem a
Thanks for the tip Ken. I tried that but don't see the importing happening
when
I check up on the status.
Below is what's in my dataimport.properties.
#Wed Nov 10 11:36:28 PST 2010
metadataObject.last_index_time=2010-09-20 11\:12\:47
interval=1
port=8080
server=localhost
params=/select?qt\=/da
Thanks for all your help Ezequiel. I cannot see anything in my email
that would make this get marked as spam. Anybody have any ideas on how
to get this fixed so I can email my questions?
robo
On Wed, Nov 10, 2010 at 11:36 AM, Ezequiel Calderara wrote:
> Tried to forward the mail of robomon but
Tried to forward the mail of robomon but had the same error:
Delivery to the following recipient failed permanently:
solr-user@lucene.apache.org
Technical details of permanent failure:
Google tried to deliver your message, but it was rejected by the recipient
domain. We recommend contacting the
Hi,
I'm not sure this is the right mail to write to, hopefully you can help or
direct me to the right person
I'm using solr - one master with 17 slaves in the server and using solrj as the
java client
Currently there's only one core in all of them (master and slaves) - only the
cpaCore.
I th
Hi,
I had a few questions regarding Solr.
Say my schema file looks like
and i index data on the basis of these fields. Now, incase i need to add a
new field, is there a way i can add the field without corrupting the
previous data. Is there any feature which adds a new field with a default
va
Mmmm maybe its your mail address? :P
Weird, i didn't have any problem with it using gmail...
Send in plain text, avoid links or links... maybe that could work...
If you want, send me the mail and i will forward it to the list, just to
test!
On Wed, Nov 10, 2010 at 3:59 PM, robo - wrote:
> No m
Sorry, OT but its driving me nuts.
I've indexed a document with chinese characters in its title. When I
perform the search (that returns json) I get back the title and using
Javascript place it into a variable that ultimately ends up as a
dropdown of titles to choose from. The problem is the
I don't understand, when the user search for perequê you want the results for
perequê and pereque?
If thats the case, any field type with ISOLatin1AccentFilterFactory should
work.
The accent should be removed at index time and at query time (Make sure the
filter is being applied on both cases)
No matter how much I limit my other email it will not get through the
Solr mailing spam filter. This has to be the most frustrating mailing
list I have ever tried to work with. All I need are some answers on
replication and load balancing but I can't even get it to the list.
On Wed, Nov 10, 201
: ExternalFileField can only be used for boosting. It is not a
: "first-class" field.
correct, but it can be used in a FunctionQuery, which means it can be
filtered on (using "frange") and (on trunk) it can be sorted on, which are
the two needs the OP asked about...
: > : Another approach woul
On Wed, Nov 10, 2010 at 1:11 PM, robo - wrote:
> How do people email this list without getting spam filter problems?
>
Depends on which side of the spam filter that you're referring to.
I've found that to keep these emails from entering my spam filter is
to add a rule to Gmail that says "Never se
I was reading the in Solr Wiki about creating request handlers -
http://wiki.apache.org/solr/SolrRequestHandler and saw that there are two
different ways to create a handler:
1. Define as and call via
http://localhost:8983/baz/?..
2. Define as and call via
http://localhost:8983/select/?qt=baz.
I tried that as well but the original email I was trying to send about
replication and load balancing was still being marked as spam (5.8 is
above threshold). That is when I thought I would try a very simple
email such as this one.
Is there a list of keywords to avoid?
On Wed, Nov 10, 2010 at 10
Hi robo,
try to send eMail in plain-text format.
This often helps a lot!
Regards
--
View this message in context:
http://lucene.472066.n3.nabble.com/Best-practice-for-emailing-this-list-tp1877693p1877792.html
Sent from the Solr - User mailing list archive at Nabble.com.
: References: <4cd8fb5a.9040...@srce.hr> <001701cb7fe2$58abc660$0a0353...@com>
:
: <4cd962b0.2090...@srce.hr>
:
: <4cd9ae1e.8080...@srce.hr>
:
: Subject: scheduling imports and heartbeats
: In-Reply-To:
http://people.apache.org/~hossman/#threadhijack
Thread Hijacking on Ma
On 11/10/2010 11:00 AM, Skreo wrote:
I just made a test :
avg-cpu: %user %nice %system %iowait %steal %idle
2,230,001,40 28,110,00 68,26
Your iowait percentage for that 10 second interval was 28%, which is
pretty high. Solr has to make a complete copy of
Hi all,
Somebody knows how can I config my solr to make searches with and without
accents?
for example:
pereque and perequê
When I do it I need the same result, but its not working.
tks
--
How do people email this list without getting spam filter problems?
I just made a test :
Before the optimize :
# w
sk...@gedeon:~$ w
18:55:43 up 22 days, 22:27, 4 users, load average: 0,07, 0,02, 0,00
USER TTY FROM LOGIN@ IDLE JCPU PCPU WHAT
# iostat 10
avg-cpu: %user %nice %system %iowait %steal %idle
0,820,00
On 2010-11-10 18:08, Skreo wrote:
Hello,
I've a Solr index with few millions of entries, so an "optimize" usually
takes several minutes (or at least tens of seconds). During the optimize
process, there is free RAM and the load average stays under 1.
But... the machine also hosts a MySQL server,
PS: There's also, I think, a way to turn on HTTP-level caching for Solr,
which I believe is caching of entire responses that match an exact Solr
query, filled without actually touching Solr at all. But I'm not sure
about this, because I'm always trying to make sure this HTTP-level cache
is turn
You know, on further reflection, I'd suggest you think (and ideally
measure) hard about whether you even need this application-level
solr-data-cache.
Solr is a caching machine, it's kind of what Solr does, one of the main
focuses of Solr. A query to Solr that hits the right caches comes back
Hello,
I've a Solr index with few millions of entries, so an "optimize" usually
takes several minutes (or at least tens of seconds). During the optimize
process, there is free RAM and the load average stays under 1.
But... the machine also hosts a MySQL server, and during the optimize
process, th
I know this post was a while ago, but it's the only one I've found which
exactly matches what we're seeing with our solr application. We recently
upgraded to 1.4.1 and all of the issues you have listed are happening to us.
Did you find a solution? Thanks.
--
View this message in context:
http
Jonathan,
sound like it makes sense.
In this case I think it is more important to size the external cache very
well, instead of Solr's.
Even when 1/5th of the requests are redundant, an external cache could not
answer the other 4/5ths and so decreasing Solr's cache would slow down the
whole appl
Shalin / Ken,
Thanks a lot for your suggestions ..I havent tried NGrams filter.. I will
try that too..
Thanks,
Barani
--
View this message in context:
http://lucene.472066.n3.nabble.com/spell-check-vs-terms-component-tp1870214p1877233.html
Sent from the Solr - User mailing list archive at Nab
Hi,
I have a peculiar situation where we are trying to use SOLR for indexing
multiple tables (There is no relation between these tables). We are trying
to use the SOLR index instead of using the source tables and hence we are
trying to create the SOLR index as that of source tables.
There are 3
Ahh this reconfirms. The analyzers are properly pulling things apart. There
are two instances of the query keyword with words between them. But from
your last comment, it sounds like the system's not trying to do any sort of
phrase highlighting, but is just hitting a weird edge case? I'm seeing thi
Em wrote:
My first query is "red firefox" - all caches were turned on.
If I am searching now for "red star", does this query makes any usage from
the cache, since both share the term "red"?
I don't believe it does, no.
I understand your question -- if your caching things externally anyway,
On Wed, Nov 10, 2010 at 7:51 AM, Em wrote:
>
> Thank you Shalin.
> Yes, both - Solr and some other applications could possible run on the same
> box.
> I hoped that not storing redundantly in Solr and somewhere else in the RAM
> would not touch Solr's performance very much.
>
> Just to understand
Thank you, James. I was looking for something like that (and I remember
having stumbled over it, in the past, now that you mention it).
I've created an xslt file that transforms the regular result to an
update xml document. Seeing that the SolrEntityProcessor is still in
development, I will stick
Thank you Shalin.
Yes, both - Solr and some other applications could possible run on the same
box.
I hoped that not storing redundantly in Solr and somewhere else in the RAM
would not touch Solr's performance very much.
Just to understand Solr'c caching mechanism:
My first query is "red firefox
On Tue, Nov 9, 2010 at 6:00 PM, Adam Estrada
wrote:
> Thanks a lot for all the tips, guys! I think that we may explore both
> options just to see what happens. I'm sure that scalability will be a huge
> mess with the core-per-user scenario. I like the idea of creating a user ID
> field and agree t
Hi all,
I seem to be lost in the new flex indexing api. In the older api I
was able to extend QueryComponent with my custom component, parse a
restricted-syntax user query into a SpanQuery, and then grab an
IndexReader. From there I worked with the spanquery's spans. For a bit
of reference
On Wed, Nov 10, 2010 at 2:57 AM, Em wrote:
>
> Hi List,
>
> in one of our application's use-case scenarios we create a response from
> different data sources.
> In clear words: We combine different responses from different data sources
> (SQL, another Webservice and Solr) to one response.
>
> We
Hi Christopher,
I am working my way through trying to implement SpanQueries in Solr
(svn trunk). From my lack of progress, I am skeptical that I can help
much, but I would be happy to try.
I imagine you have already found (either before your message, or
after posting it) Grant's lucen
I'm not sure, but SOLR-1499 might have what you want.
https://issues.apache.org/jira/browse/SOLR-1499
James Dyer
E-Commerce Systems
Ingram Content Group
(615) 213-4311
-Original Message-
From: Chantal Ackermann [mailto:chantal.ackerm...@btelligent.de]
Sent: Wednesday, November 10, 2010
On 07.11.2010, at 20:13, Lukas Kahwe Smith wrote:
> Hi,
>
> I am pondering making use of field collapsing. I am currently indexing
> clauses (sections) inside UN documents:
> http://resolutionfinder.org/search/unifiedResults?q=africa&=&t[22]=medication&dc=&st=clause
>
> Now since right now my
Hi,
If your index is supposed to handle only public information, i.e. public RSS
feeds, then I don't see a need for multiple cores.
I would probably try to handle this on the query side only. Imagine this
scenario:
User A registers RSS-X and RSS-Y (the application starts pulling and indexing
Ah I see: like you said it's part of the facet range implementation.
Frontend is already working, just need the 'update-on-slide' behavior.
Thanks
Geert-Jan
2010/11/10 gwk
> On 11/9/2010 7:32 PM, Geert-Jan Brits wrote:
>
>> when you drag the sliders , an update of how many results would match i
i'm looking for another solution other than cron job.
can i configure solr to schedule imports?
From: Ranveer Kumar
To: solr-user@lucene.apache.org
Sent: Tue, November 9, 2010 8:13:03 PM
Subject: Re: scheduling imports and heartbeats
You should use cron for t
On 11/9/2010 7:32 PM, Geert-Jan Brits wrote:
when you drag the sliders , an update of how many results would match is
immediately shown. I really like this. How did you do this? IS this
out-of-the-box available with the suggested Facet_by_range patch?
Hi,
With the range facets you get the face
Hi List,
in one of our application's use-case scenarios we create a response from
different data sources.
In clear words: We combine different responses from different data sources
(SQL, another Webservice and Solr) to one response.
We would cache this information per request for a couple of min
Dear all,
my use case is:
Creating an index using DIH where the sub-entity is querying another
SOLR index for more fields.
As there is a very convenient attribute "useSolrAddSchema" that would
spare me to list all the fields I want to add from the other index, I'm
looking for a way to get the sea
On Tue, Nov 9, 2010 at 10:16 PM, Tri Nguyen wrote:
> Hi,
>
> Can I configure solr to schedule imports at a specified time (say once a day,
> once an hour, etc)?
>
> Also, does solr have some sort of heartbeat mechanism?
>
> Thanks,
>
> Tri
Tri,
If you use the DataImportHandler (DIH), you can set
81 matches
Mail list logo