I would additionally suggest to use embeddedSolrServer for large uploads if
possible, performance are better.
2010/5/31 Steve Kuo
> I have a newbie question on what is the best way to batch add/commit a
> large
> collection of document data via solrj. My first attempt was to write a
> multi-th
Can some body throw some ideas, on how to achieve (interleaving) from with in
the application especially in a distributed setup?
“ There are only 10 types of people in this world:-
Those who understand binary and those who don’t “
Regards,
P.N.Raju,
From
I have some experience using MLT with the StandardRequestHandler with Python
but I can't figure out how to do it with solrj. It seems that to do
MLT with solrj I have
to use MoreLikeThisRequestHandler and there seems no way to use
StandardRequestHandler for MLT with solrj (please correct me if I'm
Hi Chris,
Thanks for your insights. I totally understand your point about steps 4 and
5. I wanted to control the moment when the swap would happen on the slave
side but as you say there is no use for that. It only adds up complexity
that internal solr mechanisms are already providing.
For the
Thinking twice about this architecture ...
I'm concerned about the way I'm going to automate the following steps:
A- The slaves would regularly poll Master-core1 for changes
B- A backup of the current index would be created
C- Re-Indexing will happen on Master-core2
D- When Indexing is done, we
Hello..
We have about 4 Million Products in our our Database and the Import takes
about 1,5 hours. In this Time is the Performance of the Database very bad
and our Server crashed sometimes. It's seems that DIH send only ONE select
to the db ?!?! is that right ?
all other processes cannot connec
Hi,
I read over the SpatialWiki. It wasn't clear how to query for
documents with LatLon fields
that reside inside a specific bounding box (not distance from). Simply
put, I have a google map
and want to construct a query for single LatLon fields that are inside
the map view (between the lat/lon c
Indeed, it's just a matter of ordening the results on the client-side IFF I
infer correctly from your description that you are guarenteed to get results
from enough different customers from SOlr in the first place to do the
interleaving that you describe. (In general this is a pretty big IF).
So a
do you think that the option
responseBuffer="adaptive"
should solve my problem ?
>From DIH FAQ ...:
I'm using DataImportHandler with MS SQL Server database with sqljdbc driver.
DataImportHandler is going out of memory. I tried adjustng the batchSize
values but they don't seem to make any
another question
we have sometimes a load from over 3,.. on our server and only from
different tomcat instances. no import is running and not much requests send
to solr.
we have 4 cores running for our search.
2 cores have each 4 Million doc's and the other two cores have each around
200.
Did anybody find a way to fix this more than removing the
HTMLStripCharFilter analyzer during the indexing?
Thanks
On Sat, Mar 13, 2010 at 7:55 PM, Lance Norskog wrote:
> HTMLStripCharFilter is only in the analyzer: it creates searchable
> terms from the HTML input. The raw HTML is stored and f
Hi,
Now we are getting the following exception [1] under
admin/replication/index.jsp and I have no clue what the cause could be
and couldn't find further info about it...
And how can I configure that the indices log into different log files
under the multi-index setup for tomcat [2]?
Regards,
Pe
Hi,
In order to use the current trunk version of solr, I built it running
"ant package" in trunk/solr and added the resulting maven artifacts to
my project.
Unfortunately the generated pom.xml-files contain the dependency to
lucene-*-2.9.1, but are compiled with the contents of
trunk/solr/lucene-
Hi,
I have a use case where I have to highlight indexed field values in
arbitrary text without indexing arbitrary text.
Here is an example
*Indexed field values are*
Lucid
Platform
Solr
*Arbitrary text (not indexed)*
Lucid Imagination and Guardian News and Media today announced that the
Guardia
Hi,
I have this use case: I update index every 10 min on a master-solr (via batch)
and replicate them to slaves. The clients use the slaves. From client view now
it's ugly: it looks like we change our index only every 10 minutes. Sure, but
idea now is to index all documents with a index date, s
> I have some experience using MLT with
> the StandardRequestHandler with Python
> but I can't figure out how to do it with solrj. It seems
> that to do
> MLT with solrj I have
> to use MoreLikeThisRequestHandler and there seems no way to
> use
> StandardRequestHandler for MLT with solrj (please co
hello.
this ist my request to solr. and i cannot change this.:
http://host/solr/select/?q=string
i cannot change this =( so i have a new termsComponent. i want to use
q=string as default for terms.prefix=string.
can i do somethin like this: ?
true
suggest
index
${???}
The settings and defaults will depend on which version of SQL Server you are
using and which version of the JDBC driver.
The default for resonseBuffering was changed to adaptive after version 1.2
so unless you are using 1.2 or earlier you don't need to set it to adaptive.
Also if I remember corre
thx for the reply =)
i try out selectMethod="cursor" but the load of the server is going bigger
and bigger during a import =(
selectMethod="cursor" only solve the problem with the locking ? right ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/DIH-Full-Import-DB-and-Pe
Performance is dependent on your server/data and the batchsize. To reduce
the server load experiment with different batchsize settings. The higher the
batch size the faster the import and the higher your SQL Server load will
be. Try starting with a small batch and then gradually increasing it.
Col
: So it seems like Luke does not understand Solr's long type. This
: is not a native Lucene type?
No, Lucene has concept of "types" ... there are utilities to help encode
some data in special ways (particularly numbers) but the underlying lucene
index doesn't keep track of when/how you do ths
:
: I have a suggestion for improving relevance functions in Solr by way of
: providing access to a set of pre-defined constants in Solr queries.
: Specifically, the number of documents indexed, the number of unique terms in
: a field, the total number of terms in a field, etc. are some of the
: q
On Tue, Jun 1, 2010 at 11:57 AM, Chris Hostetter
wrote:
> :
> : I have a suggestion for improving relevance functions in Solr by way of
> : providing access to a set of pre-defined constants in Solr queries.
> : Specifically, the number of documents indexed, the number of unique terms
> in
> : a f
The insert shards code is as follows:
ModifiableSolrParams modParams = new ModifiableSolrParams(params);
modParams.set("shards", shards);
rb.req.setParams(modParams);
Where shards is a valid single shard pseudo URL.
Stacktrace:
HTTP Status 500 - null java.lang.NullPointerException at
org.apache
Thanks, the pointer to that documentation page (which somehow I had
missed), as well as Chris's response is very helpful.
The one thing I'm still not sure about, which I might be able to figure
it out through trial-and-error reverse engineering, is escaping issues
when you combine nested querr
It's not clear if you're just trying to figure it all out, or get
something specific to work.
If you can give a specific example, we might be able to suggest easier
ways to achieve it rather than going escape crazy :-)
-Yonik
http://www.lucidimagination.com
On Tue, Jun 1, 2010 at 5:06 PM, Jonat
:
http://lucene.472066.n3.nabble.com/StackOverflowError-during-Delta-Import-td811053.html#a824780
yeah, i remember that thread -- it really seems like a driver issue, but
understandable that "fixing hte driver" is probably more out of scope then
"working arround in solr"
: I never did find a
:
: where can I find more information about a failure of a Java replication
: in Solr 1.4?
: (Dashboard does not seem to be the best place!?)
All the log message are written using the JDK Logging framework, so it
really depends on your servlet container, and where it's configured to
write the l
You have to give us more details then that if you expect anyone to have
a clue what might be going wrong...
* what does your code for initializing solr look like?
* what does your soler home dir look like (ie: what files are in it)
* what do all of your config files look like?
* what is the full
On 01.06.2010, at 23:35, Chris Hostetter wrote:
>
> :
> http://lucene.472066.n3.nabble.com/StackOverflowError-during-Delta-Import-td811053.html#a824780
>
> yeah, i remember that thread -- it really seems like a driver issue, but
> understandable that "fixing hte driver" is probably more out o
: CommonsHttpSolrServer.request() resulting in multiple searchers. My first
: thought was to change the configs for autowarming. But after looking at the
: autowarm params, I am not sure what can be changed or perhaps a different
: approach is recommened.
even with 0 autowarming (which is what
: In order to use the current trunk version of solr, I built it running
: "ant package" in trunk/solr and added the resulting maven artifacts to
: my project.
the trunk is definitley in flux now with the way Lucene & solr (and the
new "modules" directory) are all designed to be built as one mono
I am just trying to figure it out mostly, the particular thing I am
trying to do is a very general purpose mapper to complex dismax nested
querries. I could try to explain it, and we could go back and forth for
a while, and maybe I could convince you it makes sense to do what I'm
trying to do.
Wild shot in the dark: if the list of shards is changed between prepare
method and hte process method of the QueryComponent that could tickle some
code path that was never expected, and maybe trigger an NPE (ie: looking
up some IDs in a map keyed off of shard and now the shard is something
tha
I'll give the deletedEntity "trick" a try... igneous
--
View this message in context:
http://lucene.472066.n3.nabble.com/Subclassing-DIH-tp830954p863108.html
Sent from the Solr - User mailing list archive at Nabble.com.
: Any idea why this query returns 0 records:
: "sexual assault" AND (-obama)
: while this one returns 1400 ?
: "sexual assault" AND -(obama)
in the first one, the parans create a boolean query consisting of a single
negated clause -- but pure negative boolean queries (ie: boolean queries
Hmmm, well, the lucene query parser does basic backslash escaping, and
so does local params within quoted strings. You can also use
parameter derefererencing to avoid the need to escape values too.
Like you pointed out, using single quotes in some places can also
help.
But instead of me trying to
This seems to be a problem (from my limited understanding). I
encountered the same thing.
And the problem is that you can have results that independently match
the constraints for latitude and longitude, but the corresponding points
would not match (i.e. with separate fields there are no longer po
Sorry for the noise, but thought I would send out a reminder to get your talks
in...
On May 17, 2010, at 8:43 AM, Grant Ingersoll wrote:
> Lucene Revolution Call For Participation - Boston, Massachusetts October 7 &
> 8, 2010
>
> The first US conference dedicated to Apache Lucene and Solr is
In the "/spell" declaration in the example solrconfig.xml, we find
these lines among the default parameters:
spellcheck
How does one supply such an array of strings in HTTP parameters? Does
Solr have a parsing option for this?
--
Lance Norskog
goks...@gmail.com
We have around 5 million items in our index and each item has a description
located on a separate physical database. These item descriptions vary in
size and for the most part are quite large. Currently we are only indexing
items and not their corresponding description and a full import takes arou
Thanks a lot for the quick responses. I will try it out.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Generic-question-on-Query-Analyzers-tp849075p863512.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi All,
When I query for a word say Tiger woods, and sort results by score... i do
notice that the results are mixed up i.e first 5 results match Tiger woods
the next 2 match either tiger/tigers or wood/woods
the next 2 after that i notice again match tiger woods.
How do i make sure that when se
One way to do it would be to use dismax request handler at query time, with a
pf paramater on the same field(s) as your qf paramter, but with a big boost on
the pf. http://wiki.apache.org/solr/DisMaxRequestHandler
I'm not sure why you're getting matches for "tigers" and "woods" on "tiger
woods
That's it. Thank you!
I thought mlt was available by default. I was wrong.
On Tue, Jun 1, 2010 at 8:22 AM, Ahmet Arslan wrote:
>> I have some experience using MLT with
>> the StandardRequestHandler with Python
>> but I can't figure out how to do it with solrj. It seems
>> that to do
>> MLT with s
45 matches
Mail list logo