We were having performance issues using servers running on VM. Are you
running QA or Prod in a VM?
2009/6/21, Stephen Weiss :
> Isn't it possible that the production equipment is simply under much
> higher load (given that, since it's in production, your various users
> are all actually using it),
Can DIH read item by item instead of the whole file before indexing? my biggest
file size is 6GB, larger than the JVM max ram value.
--- On Sat, 6/20/09, Erik Hatcher wrote:
> From: Erik Hatcher
> Subject: Re: Use DIH with large xml file
> To: solr-user@lucene.apache.org
> Date: Saturday, Ju
Isn't it possible that the production equipment is simply under much
higher load (given that, since it's in production, your various users
are all actually using it), vs the QA equipment, which is only in use
by the people doing QA?
We've found the same thing at one point - we had a very sm
Hi Francis,
I can't tell what the problem is from the information you've provided so far.
My gut instinct is that this is due to some difference in QA vs. PROD
environments that isn't Solr-specific.
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Message --
Yeah the lucid imagination articles are great!
Jonathan, you can also use the dismax query parser and apply boosts using
the qf (query fields) param:
q=my query here&qf=title^0.5 author^0.1
http://wiki.apache.org/solr/DisMaxRequestHandler#head-af452050ee272a1c88e2ff89dc0012049e69e180
Matt
On S
Hi Jonathan,
I think this is the best article related to faceted search.
http://www.lucidimagination.com/Community/Hear-from-the-Experts/Articles/Faceted-Search-Solr
On Sat, Jun 20, 2009 at 9:56 PM, Jonathan Vanasco wrote:
> i've gone through the official docs a few times, and then found some
How are you configuring DIH to read those files? It is likely that
you'll need at least as much RAM to the JVM as the largest file you're
processing, though that depends entirely on how the file is being
processed.
Erik
On Jun 20, 2009, at 9:23 PM, Jianbin Dai wrote:
Hi,
I hav
Hi,
I have about 50GB of data to be indexed each day using DIH. Some of the files
are as large as 6GB. I set the JVM Xmx to be 3GB, but the DIH crashes on those
big files. Is there any way to handle it?
Thanks.
JB
i've gone through the official docs a few times, and then found some
offsite stuff of varying quality regarding how-tos.
can anyone here recommend either howtos/tutorials or sample
applications that they have found worthwhile ?
specifically i'm looking to do the following:
- with regular
Thank you Erik and Otis!
I tried the switch to string and it worked perfectly!
Int wasn't appropriate, because our UIDs are based on object-type +
object-id, and i'd rather not work on some method of creating
universal serials for everything.
On Jun 19, 2009, at 11:58 PM, Erik Hatcher wrot
https://issues.apache.org/jira/browse/SOLR-1191
describes a different problem but I think his Ali's solution applies here.
I tried 'select concat("",id) from table' and this also had the same
exception. I can't test now, but I think this is the solution:
select concat("prefix",id) AS ID
The JDB
Do you have a default field declared? &ext.default.fl=
Either that, or you need to explicitly capture the fields you are
interested in using &ext.capture=
You could add this to your curl statement to try out.
-Grant
On Jun 20, 2009, at 8:41 AM, cloax wrote:
Thanks for the quick respon
Thanks for the quick response.
Here are the fields from the schema:
I use text as the content field for the default field for the ERH.
Here's the config of the ERH:
last_modified
true
Here's the output of a curl request w/ the file:
0650
Thanks,
The result of adding &debugQuery=true follows. Does this mean the order is
always defined from score? If that is the case, do I have to adjust the way
how solr calculate score? How can I do that?
I followed Otis's sugestion also and add &sort=word+asc but the first one is
not "apple", bu
Hello,
I am indexing books.
Fields are Title,Author, Subtitle, Category,Pages ...
Boosts are TitleBoost 1, AuthorBoost .8 SubtitleBoost .6;
Some one entered a query to search lets say Hitchhiker Guide
Now I want to show the reuslts in which both the words occur based on there
boost score.
I
15 matches
Mail list logo