I have a dismax query where I check for values in 3 fields against documents in
the index - a title, a list of keyword tags and then full-text of the document.
I usually get lots of results and I can see that the first results are OK -
it's giving precedence to titles and tag matches, as my dism
Hi all,
Having read the SolrPerformanceFactors wiki page [1], I'd still need a
couple of clarifications about mergeFactor (I am using version 1.4.1) so if
anyone can help it would be nice.
- Is mergeFactor a one time configuration setting that is considered only
when creating the index for t
Hi,
I've worked around the issue by setting omitNorms=true on the title field. Now
all fieldNorm values are 1.0f and therefore do not mess up my scores anymore.
This, of course, is hardly a solution even though i currently do not use
index-time boosts on any field.
The question remains, why do
Hi!
I am trying to get the ContentStreamDataSource to work properly , but there are
not many examples out there.
What I have done is that I have made a copy of my HttpDataSource config and
replaced the
Hi all,
having two fields named 'type' and 'cat' with identical type and options,
but different values recorded, would it be possible to query them as they
were one field?
For instance
q=type:electronics cat:electronics
should return same results as
q=common:electronics
I know I could make it def
On Thu, Nov 4, 2010 at 8:21 AM, Tommaso Teofili
wrote:
> Hi all,
> having two fields named 'type' and 'cat' with identical type and options,
> but different values recorded, would it be possible to query them as they
> were one field?
> For instance
> q=type:electronics cat:electronics
> should r
Ken's suggestion to look at dismax is a good one, but I have
a question
q=type:electronics cat:electronics
should do what you want assuming your default operator
is OR. Is it failing? Or is the real question how you can
do this automatically?
I'd expect the ranking to be a bit different, but I'm
for contentstreamdatasource to work you must post the stream in the request
On Thu, Nov 4, 2010 at 8:13 AM, Theodor Tolstoy
wrote:
> Hi!
> I am trying to get the ContentStreamDataSource to work properly , but there
> are not many examples out there.
>
> What I have done is that I have made a cop
On 11/4/2010 3:27 AM, Tommaso Teofili wrote:
- Is mergeFactor a one time configuration setting that is considered only
when creating the index for the first time or can it be adjusted later even
with some docs inside the index? e.g. I have mF to 10 then I realize I want
quicker se
On Thu, Nov 4, 2010 at 8:04 AM, Markus Jelsma
wrote:
> The question remains, why does the title field return a fieldNorm=0 for many
> queries?
Because the index-time boost was set to 0 when the doc was indexed. I
can't say how that happened... look to your indexing code.
> And a subquestion, do
Have you tried encoding it with %3F?
firstname:*%3F*
On 2010-11-04, at 1:44 AM, Stephen Powis wrote:
> I'm having difficulty properly escaping ? in my search queries. It seems as
> tho it matches any character.
>
> Some info, a simplified schema and query to explain the issue I'm having.
> I'm
On Thu, Nov 4, 2010 at 1:44 AM, Stephen Powis wrote:
> I want to return any first name with a Question Mark in it
> Query: first_name: *\?*
>
There is no way to escape the metacharacters * or ? for a wildcard
query (regardless of queryparser, even if you write your own).
See https://issues.apache
I've done some testing with the example docs and it behaves similar when there
is a zero doc boost. Luke, however, does not show me the index-time boosts.
Bost document and field boosts are not visible in Luke's output. I've changed
doc boost and field boosts for the mp500.xml document but all i
On Thu, Nov 4, 2010 at 9:51 AM, Markus Jelsma
wrote:
> I've done some testing with the example docs and it behaves similar when there
> is a zero doc boost. Luke, however, does not show me the index-time boosts.
Remember that the norm is a product of the length norm and the index
time boost... it
On Thursday 04 November 2010 15:12:23 Yonik Seeley wrote:
> On Thu, Nov 4, 2010 at 9:51 AM, Markus Jelsma
>
> wrote:
> > I've done some testing with the example docs and it behaves similar when
> > there is a zero doc boost. Luke, however, does not show me the
> > index-time boosts.
>
> Remember
Hi,
I have data stored in a database that is being updated constantly. I need to
find a way to update Solr index as data in the database is being updated.
There seems to be 2 main schools of thoughts on this:
1) DIH delta - query the database for all records that have a timestamp later
than the l
I'm in the same scenario, so this answer would be helpful too..
I'm adding...
3) Web Service - Request a webservice for all the new data that has been
updated (can this be done?
On Thu, Nov 4, 2010 at 2:38 PM, Andy wrote:
> Hi,
> I have data stored in a database that is being updated constantly.
On 11/4/2010 7:22 AM, stockiii wrote:
how can i start an optimize by using DIH, but NOT after an delta- or
full-import ?
I'm not aware of a way to do this with DIH, though there might be
something I'm not aware of. You can do it with an HTTP POST. Here's
how to do it with curl:
/usr/bin/c
Hi all-
First, thanks to all the folks to have helped me so far getting the hang of
Solr; I promise to give back when I think my contributions will be useful :)
I am at the point where I'm trying to return results back from a search in a
war file, using Java with solrj. On the result page of th
For what it's worth, the Solr class instructor at the Lucene Revolution
conference recommended *against* optimizing, and instead suggested to just
let the merge factor do it's job.
On Thu, Nov 4, 2010 at 2:55 PM, Shawn Heisey wrote:
> On 11/4/2010 7:22 AM, stockiii wrote:
>
>> how can i start an
Huh? That's something new for me. Optmize removed documents that have been
flagged for deletion. For relevancy it's important those are removed because
document frequencies are not updated for deletes.
Did i miss something?
> For what it's worth, the Solr class instructor at the Lucene Revoluti
I need to connect to a RETS api through a http url. But the REST service uses
digest authentication. Can I use DataImportHandler to pass the credentials
for digest authentication?
Thanks
--
View this message in context:
http://lucene.472066.n3.nabble.com/Does-DataImportHandler-support-Digest-aut
I mean to say RESTful Apis.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Does-DataImportHandler-support-Digest-authentication-tp1844497p1844501.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I'm now trying to
export JAVA_OPTS="$JAVA_OPTS -Dsolr.solr.home=/path/to/my/schema.xml"
and restarting tomcat (v6 package from ubuntu maverick) via
sudo /etc/init.d/tomcat6 restart
but solr still doesn't seem to find that schema.xml, as it complains
about unknown fields when running the
Looking at the JIRA issue, looks like there's been a new patch related to
this. This is good news! We've re-written a portion of our web app to use
Solr instead of mysql. This part of our app allows clients to construct
rules to match data within their account, and automatically apply actions to
I believe it should point to the directory above, where conf and lib are
located (though I have a multi-core setup).
Mine is set to:
/usr/local/jboss-5.1.0.GA/server/solr/solr_data/
And in solr_data the solr.xml defines the two cores, but in each core
directory, is a conf, data, and lib direct
I'm still seeing this error after downloading the latest 2.9 branch
version, compiling, copying to Solr 1.4 and deploying. Basically as
mentioned, the .del files are of zero length... Hmm...
On Wed, Oct 13, 2010 at 1:33 PM, Jason Rutherglen
wrote:
> Thanks Robert, that Jira issue aptly describes
Wildcard queries, especially a wildcard query with a wildcard both
_before_ and _after_, are going to be fairly slow for Solr to process,
anyhow. (In fact, for some reason I thought wildcards weren't even
supported both before and after, just one or the other).
Still, it's a bug in lucene, it
You need to either add that to catalina.sh or create a setenv.sh in the
CATALINA_HOME/bin directory. Then you can restart tomcat.
So, setenv.sh would contain the following:
export JAVA_HOME="/path/to/jre"
export JAVA_OPTS="="$JAVA_OPTS -Dsolr.solr.home=/path/to/my/schema.xml"
If you we
Hi Ron,
how do I know what the starting row
Always 0.
especially if the original SolrQuery object has them all
thats the point. solr will normally cache it for you. This is your friend:
40
just try it first with http to get an impression what start is good for:
it just sets the starti
what you can try maxSegments=2 or more as a 'partial' optimize:
"If the index is so large that optimizes are taking longer than desired
or using more disk space during optimization than you can spare,
consider adding the maxSegments parameter to the optimize command. In
the XML message, this
The thing is, I only have a schema.xml -- no data, no lib directories.
See the tests subdirectory in the solrpy package:
http://pypi.python.org/packages/source/s/solrpy/solrpy-0.9.3.tar.gz
Bernhard
Am Donnerstag, den 04.11.2010, 15:59 -0500 schrieb Olson, Ron:
> I believe it should point to the
On Thu, Nov 4, 2010 at 4:58 PM, Stephen Powis wrote:
> What is the likelihood of this being included in the next release/bug fix
> version of Solr?
In this case, not likely. It will have to wait for Solr 4.0
> Are there docs available online with basic information
> about rolling our own build o
Thanks for your instructions. Unfortunately, I need to do all that as
part of my package's (python-solrpy) build procedure, so I can't change
any global configuration, such as in the catalina subdirectories.
I've already sensed that restarting tomcat is also just too
system-invasive and would incl
Hi Erick
2010/11/4 Erick Erickson
> Ken's suggestion to look at dismax is a good one, but I have
> a question
> q=type:electronics cat:electronics
>
> should do what you want assuming your default operator
> is OR.
correct
> Is it failing? Or is the real question how you can
> do this autom
You can setup your own tomcat instance which would contain just configurations
you need. You won't even have to recreate all the tomcat configuration and
binaries, just the ones that were not defaults. So, if you lookup multiple
tomcat configuration instance (google it), and then you'll have a
Tommaso Teofili wrote:
No failing, just looking for how to do such "expansion" of fields
automatically (with fields in OR but that's not an issue I think)
the dismax query parser is that way.
Thanks so much Shawn, I am in a scenario with many inserts while searching,
each consisting of ~ 500documents, I will monitor the number of
segments taking your considerations in mind :-)
Regards,
Tommaso
2010/11/4 Shawn Heisey
> On 11/4/2010 3:27 AM, Tommaso Teofili wrote:
>
>>- Is mergeFa
Hi Jayant,
I think you mean NL search as opposed to Boolean search: the ability to return
ranked results from queries based on non-required term matches. Right?
If that is what you meant, then the answer is: "Yes!". If not, then you should
rephrase your question.
Otherwise, the answer coul
Hi,
don't know if the python package provides one but solrj offers to start
solr embedded (|EmbeddedSolrServer|) and
setting up different schema + config is possible. for this see:
https://karussell.wordpress.com/2010/06/10/how-to-test-apache-solrj/
if you need an 'external solr' (via jetty a
Hi all,
we had a severe problem with our raidcontroller on one of our servers today
during importing a table with ~8 million rows into a solr index. After
importing about 4 million
documents, our server shutdown, and failed to restart due to a corrupt raid
disk.
The Solr data import was the on
no, you didn't miss anything. The comment at Lucen Revolution was more
along the lines that optimize didn't actually improve much #absent# deletes.
Plus, on a significant size corpus, the doc frequencies won't changed that
much by deleting documents, but that's a case-by-case thing
Best
Erick
On
I experienced similar problems. It was because we didn't perform load stress
tests properly, before going to production. Nothing is forever, replace
controller, change hardware vendor, maintain low temperature inside a rack.
Thanks
--Original Message--
From: Robert Gründler
To: solr-user
I don't think current lucene will offer what you want now.
There are 2 main tasks in a search process.
One is "understanding" users' intension. Because natural language
understanding is difficult, Current Information Retrival systems
"force" users input some terms to express their needs
I am able to facet on a particular field because I have index on that field.
But I am not sure how to facet on a price range when I have the exact price
in the 'price' field. Can anyone help here.
Thanks
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-Facet-on-a-price-
sometorment later
I found the reason ofsolr replication'slow speed. It's not solr's problem.It's
jetty's. I used to embed jetty7 in my app. But when I found solr's demo use
jetty6 , I tried to use jetty6 in my app and I was so happy to get the fast
speed.
actually, I tried to change solr's d
46 matches
Mail list logo