When I first saw this particular email, I wrote a letter intend to ask the
sender remove solr-user from its recepient cause I thought this should go to
solr-dev. But then I thought again, it's about 'job-offer' not 'development
of Solr', I just delete my email.
Maybe solr-job is a good suggest
Hello all,
I'm getting stuck when trying to import oracle DB to solr index, could
any one of you give a hand. Thanks million.
Below is some short info. that might be a question
My Sorl: 1.4.1
*LOG *
INFO: Starting Full Import
Oct 29, 2010 1:19:35 PM org.apache.solr.handler.data
Hi Jakub,
I have ported the KStemmer for use in most recent Solr trunk version.
My stemmer is located in the lib directory of Solr "solr/lib/KStemmer-2.00.jar"
because it belongs to Solr.
Write it as FilterFactory and use it as Filter like:
This is how my fieldType looks like:
(Sorry - fumble finger sent too soon.)
My confusion stems from the fact that in my test I insert a number of
documents, and then retrieve them ordered by timestamp, and they don't come
back in the same order they were inserted (the order seems random), unless I
commit after each insert.
Is that
This is a great explanation, thanks. I'm going to add it to the wiki somewhere
that seems relevant, if no-one minds and the wiki lets me.
From: Chris Hostetter [hossman_luc...@fucit.org]
Sent: Thursday, October 28, 2010 7:27 PM
To: solr-user@lucene.apache
I'm curious what if any guarantees there are regarding the "timestamp" field
that's defined in the sample solr schema.xml. Just for completeness, the
definition is:
The specifics of your overall goal confuse me a bit, but drilling down to
your core question...
: I want to be able to use the dismax parser to search on both terms
: (assigning slops and tie breaks). I take it the 'fq' is a candidate for
: this,but can I add dismax capabilities to fq as well? A
: the documentCache: "(Note: This cache cannot be used as a source for
: autowarming because document IDs will change when anything in the
: index changes so they can't be used by a new searcher.)"
:
: Can anyone elaborate a bit on that. I think I've read it at least 10
: times and I'm still unab
On 10/28/2010 12:02 PM, Chris Hostetter wrote:
I'm not very knowledgeable about how distributed searching deals with
request handlers, url paths, and the qt param (i have no idea why the
exact same handler isn't propograted to the remote shards by default -- i
thought it was, but your email sugge
Pavel,
it depends on size of your documents corpus, complexity and types of
the queries you plan to use etc. I would recommend you to search for
the discussions on synonyms expansion in Lucene (index time vs. query
time tradeoffs etc.) since your problem is quite similar to that
(think Moskva vs.
hi
I added spellchecker to request handler. Spellchecker is indexed based.
Terms in index are like
iphone
iphone 4
iphone case
phone
gophoe
when i set q=iphole i get suggestions like
iphone
phone
gophone
ipad
Not sure how would i get iphone, iphone 4, iphone case, phone. Any thoughts?
At the
hi thanks.. It worked.!!
--
View this message in context:
http://lucene.472066.n3.nabble.com/spellcheck-component-does-not-work-with-request-handler-tp1786079p1789163.html
Sent from the Solr - User mailing list archive at Nabble.com.
Doing a range search is straightforward. I have a fixed value in a document
field, I search on [x TO y] and if the fixed value is in the range requested
it gets a hit. But, what if I have data in a document where there is a min
value and a max value and my query is a fixed value and I want to get
On Thu, Oct 28, 2010 at 4:44 PM, wrote:
>
> I'm using Solr 1.2. If I upgrade to 1.4.1, must I re-index because of
> LUCENE-1142? If so, how will this affect me if I don’t re-index (I'm using
> EnglishPorterFilterFactory)? What about when I’m using non-English stammers
> from Snowball?
>
> B
Hi,
I would aim for reindexing on branch3_x, which will be the 3.1 release soon. I
don't know if SOLR-1873 applies cleanly to 3_x now, but it would surely be less
effort to have it apply to 3_x than to 1.4. Perhaps you can help backport the
patch to 3_x?
--
Jan Høydahl, search solution archite
I'm using Solr 1.2. If I upgrade to 1.4.1, must I re-index because of
LUCENE-1142? If so, how will this affect me if I don’t re-index (I'm using
EnglishPorterFilterFactory)? What about when I’m using non-English stammers
from Snowball?
Beside the brief note "IMPORTANT UPGRADE NOTE" about t
Hey! I represent those remarks! I was on that committee (really) because I
am/was a:
http://www.rhyolite.com/anti-spam/you-might-be.html#spam-fighter
and about 20 other 'types' on that list. I'm a little bit more mature, but
only a little. White lists are the only way to go.
Dennis Gea
Speaking of jobs on this list . . . .
How much does a good consultant for Solr work cost?
I am interested first in English, but then in other languages around the world.
Just need budgetary amounts for a business plan.
1-6mos, or till BIG DOLLARS, whichever comes first ;-)
Dennis Gearon
Sig
http://www.rhyolite.com/anti-spam/you-might-be.html#spammers-are-stupid-3
-
SECURITY/CONFIDENTIALITY WARNING:
This message and any attachments are intended solely for the individual or
entity to which they are addressed. This
On Thu, Oct 28, 2010 at 2:57 PM, Michael McCandless <
luc...@mikemccandless.com> wrote:
> I don't think we should do this until it becomes a "real" problem.
>
> The number of job offers is tiny compared to dev emails, so far, as
> far as I can tell.
>
> Mike
>
>
By the time that it becomes a real
I don't think we should do this until it becomes a "real" problem.
The number of job offers is tiny compared to dev emails, so far, as
far as I can tell.
Mike
On Thu, Oct 28, 2010 at 2:10 PM, rajini maski wrote:
> Its better if we can make some solr-job list.. that would be better.. if
> not,
>
On Thu, Oct 28, 2010 at 5:18 PM, Michael McCandless
wrote:
> Is it somehow possible that you are trying to sort by a multi-valued field?
[...]
Either that, or or your field gets processed into multiple tokens via the
analyzer/tokenizer path in your schema. The reported error is a
consequence of t
This is why using 'groups' as intermidiary permission objects came into
existence in databases.
Dennis Gearon
Signature Warning
It is always a good idea to learn from your own mistakes. It is usually a
better idea to learn from others’ mistakes, so you do not have to make them
Its better if we can make some solr-job list.. that would be better.. if
not,
chances that this mailing list of solr queries become less of that and more
lik job forum.. this mailing list is so uselful to all developers to get
answers for their techinical queries..
On Thu, Oct 28, 2010 at 11:30
: Another approach would be to use ExternalFileField and keep the price data,
: normalized to USD, outside of the index. Every time the currency rates
: changed, we would calculate new normalized prices for every document in the
: index.
...that is the approach i would normally suggest.
: Still
Right - historically it's been fine because it hasn't grown into a
problem issue. Hopefully it just stays that way.
- Mark
On 10/28/10 2:00 PM, Stefan Moises wrote:
> Well, I don't see a problem sending (serious) job offers to this list...
> as long as nobody spams
>
> just my 2c
> Stefan
>
> A
: Is there any way to preserve qt in a distributed search so this doesn't
: happen? I am using Solr 1.4.1, but we are upgrading to 3.1-dev very soon.
I'm not very knowledgeable about how distributed searching deals with
request handlers, url paths, and the qt param (i have no idea why the
exac
Well, I don't see a problem sending (serious) job offers to this list...
as long as nobody spams
just my 2c
Stefan
Am 28.10.2010 19:57, schrieb Ravi Gidwani:
May I suggest a new mailing list like solr-jobs (if it does not exist) or
something for such emails ? I think it is also important for t
May I suggest a new mailing list like solr-jobs (if it does not exist) or
something for such emails ? I think it is also important for the solr
developers to get emails about job opportunities ? No ?
~Ravi.
On Tue, Oct 26, 2010 at 11:42 PM, Pradeep Singh wrote:
> This is the second time he has
I'm reading my document data from a CMS and indexing it using calls to
curl. The curl call includes 'stream.url' so Tika will also index the
actual document pointed to by the CMS' stored url. This works fine.
Presentation side I have a dropdown with the title of all the indexed
documents suc
On Wed, Oct 27, 2010 at 9:13 PM, Chris Hostetter
wrote:
>
> : schema.) My evidence for this is the documentCache stats reported by
> : solr/admin. If I request "rows=10&fl=id" followed by
> : "rows=10&fl=id,title" I would expect to see the 2nd request result in
> : a 2nd insert to the cache, but i
In your "standard" Search Handler, you have the "last-components" array inside
. However, it should be outside as in the "/spell" Search
Handler. Try this:
explicit
spellcheck
James Dyer
E-Commerce Systems
Ingram Content Group
(615) 213-4311
-Original Message-
From:
Hello,
We currently index our data through a SQL-DIH setup but due to our model
(and therefore sql query) becoming complex we need to index our data
programmatically. As we didn't have to deal with commit/optimise before, we
are now wondering whether there is an optimal approach to that. Is there
Hi!
There is a polish stemmer http://www.getopt.org/stempel/ and I have
problems connecting it with solr 1.4.1
Questions:
1. Where EXACTLY do I put "stemper-1.0.jar" file?
2. How do I register the file, so I can build a fieldType like:
3. Is that the right approach to make it work?
Thanks
I am using SOLR 1.3
I wanted to add spellcheck component to to standard request handler it so
did this
explicit
spellcheck
but for some reason it does not return suggestion for misspelled words. For
instance iphole does not get a suggestion of
On Thu, Oct 28, 2010 at 8:07 AM, Ephraim Ofir wrote:
> How is this better than replication?
>
> Ephraim Ofir
>
>
It's not; for our needs here, we have not set up replication through SOLR.
We are working through OOM problems/performance tuning first, then "best
practices" second. I just wanted the
How is this better than replication?
Ephraim Ofir
-Original Message-
From: Ken Stanley [mailto:doh...@gmail.com]
Sent: Thursday, October 28, 2010 1:59 PM
To: solr-user@lucene.apache.org
Subject: Re: If I want to move a core from one physical machine to another
On Wed, Oct 27, 2010
On Wed, Oct 27, 2010 at 6:12 PM, Ron Mayer wrote:
> If I want to move a core from one physical machine to another,
> is it as simple as just
> scp -r core5 otherserver:/path/on/other/server/
> and then adding
>
> on that other server's solr.xml file and restarting the server there?
>
>
>
>
Is it somehow possible that you are trying to sort by a multi-valued field?
Mike
On Thu, Oct 28, 2010 at 6:59 AM, Pablo Recio wrote:
> Hi all. I'm having a problem with solr sorting search results.
>
> When I try to make a query and sort it by title:
>
> http://localhost:8983/solr/select/?q=*%3A
Hi,
We are in a beta testing phase, with several users a day.
After several days of running well, the solr server stopped responding
to requests that require a lot of processing time, like this one :
wt=json&omitHeader=true&q=qiAndMSwFR%3A%28transport%29&q.op=AND&start=0&rows=5&fl=id,domainId
Hi all. I'm having a problem with solr sorting search results.
When I try to make a query and sort it by title:
http://localhost:8983/solr/select/?q=*%3A*&version=2.2&start=0&rows=10&indent=on&sort=title%20desc
I get that error [1]. If I try to sort by other indexed field it works, indeed
if I c
Sorry for very bad pasting. I paste it again.
Slowest Components Count Exclusive
Total
QueryElevationComponent 1 506,858 ms 100%
506,858 ms 100%
SolrIndexSearcher 1 2.0 ms
0
Hi,
I'm using solr 1.4 and using QueryElevation Component for guaranteed search
position. I have around 700,000 documents with 1 Mb elevation file. It turns
out it is quite slow on the newrelic monitoring website:
Slowest Components Count Exclusive Total QueryElevationComponent 1 506,858
ms 100
Alexander,
Thanks,
What variat has better performance?
2010/10/28 Alexander Kanarsky
> Pavel,
>
> I think there is no single way to implement this. Some ideas that
> might be helpful:
>
> 1. Consider adding additional terms while indexing. This assumes
> conversion of Russian text to both "tra
Hi all,
We are using Solr over Jetty with a large index, sharded and distributed over
multiple machines. Our queries are quite long, involving boolean and proximity
operators. We cut the connection at the client side after 5 minutes. Also, we
are using parameter timeAllowed to stop executing it
Hi,
We are in a beta testing phase, with several users a day.
After several days of waiting, the solr server didn't respond to
requests that require a lot of processing time.
I'm using Solr inside Tomcat.
This is the request that had no response from the server :
wt=json&omitHeader=true&q=q
On Wed, Oct 27, 2010 at 8:59 PM, Jonathan Rochkind wrote:
> Seriously, at least try JVM argument -XX:+UseConcMarkSweepGC . That
> argument took care of very similar symptoms I was having. I never did
> figure out exactly what was causing them, but at some point I tried that JVM
> argument, and
In our case, we just store a database id and do a secondary db query when
displaying the results.
This is handy and leads to a more centralised architecture when you need to
display properties of a domain object which you don't index/search.
On 28 October 2010 05:02, kenf_nc wrote:
>
> Interesti
is there anyone could help me?
2010/10/11 Li Li :
> hi all,
> I want to know the detail of IndexReader in SolrCore. I read a
> little codes of SolrCore. Here is my understanding, are they correct?
> Each SolrCore has many SolrIndexSearcher and keeps them in
> _searchers. and _searcher keep t
Pavel,
I think there is no single way to implement this. Some ideas that
might be helpful:
1. Consider adding additional terms while indexing. This assumes
conversion of Russian text to both "translit" and "wrong keyboard"
forms and index converted terms along with original terms (i.e. your
Analy
On Thu, Oct 28, 2010 at 3:42 AM, Ron Mayer wrote:
> If I want to move a core from one physical machine to another,
> is it as simple as just
> scp -r core5 otherserver:/path/on/other/server/
> and then adding
>
> on that other server's solr.xml file and restarting the server there?
If "core
Hi Lance,
I actually copied tika exceptions in one html file and indexed
it. It is just a content of a file and here i tell u what i mean::
if i post a query like *java* then the result or response from solr should
hit only a part of the content like as follows::
http://localhost:
52 matches
Mail list logo