Hi,
First off, we're a happy user of Apache Solr v3.1 Enterprise search server,
integrated and successfully running in our LIVE Production server.
Now, we're enhancing our existing search feature in our web application as
explained below, that truly helps application users in making informed
deci
Hi.
I would like to be able to do a facet on a date field, but with different
ranges (in a single query).
for example. I would like to show
#documents by day for the last week -
#documents by week for the last couple of months
#documents by year for the last several years.
is there a way to d
A Lucene SpanFirstQuery (with a boost) would do it, but you'd have to find a
query parse that supports it and most don't.
You could also keep a copy of the title as a string field and then use a
trailing wildcard to check if the title began with a term and boost it.
title_s:Kabul*^2.0
--
: > 1 bomb blast in kabul
: >
: > 2 kabul bomb blast
: >
: > 3 3 people killed in serial bomb blast in kabul
...
: > I want 2nd result should come first while user search by "kabul".
: >
: > Because kabul is on 1st postion in that sentance. Similarly 1st result
: > should come on 2nd and
: Is there way to perform soft commit from code in Solr 4.0 ?
: Is it possible only from solrconfig.xml through enabling autoSoftCommit
: with maxDocs and/or maxTime attributes?
http://wiki.apache.org/solr/NearRealtimeSearch
links to:
http://wiki.apache.org/solr/UpdateXmlMessages#A.22commit.22_
On 5/2/2012 5:54 AM, Prakashganesh, Prabhu wrote:
We have a fairly large scale system - about 200 million docs and fairly high
indexing activity - about 300k docs per day with peak ingestion rates of about
20 docs per sec. I want to work out what a good mergeFactor setting would be by
testing
for this version, you may consider using payload for position boost.
you can save boost values in payload.
I have used it in lucene api where anchor text should weigh more than
normal text. but I haven't used it in solr.
some searched urls:
http://wiki.apache.org/solr/Payloads
http://digitalpebble.
Make two fields, one with stores the stripped HTML and another that
stores the parsed HTML. You can use so that you do not
have to submit the html page twice.
You would mark the stripped field 'indexed=true stored=false' and the
full text field the other way around. The full text field should be
Hi Sami:
On Thu, May 3, 2012 at 12:34 AM, Sami Siren wrote:
>
> I believe that this particular change was part of
> https://issues.apache.org/jira/browse/SOLR-1730. The ability to start
> solr with no cores seems like a reasonable feature so I would classify
> this as a bug. Not sure what others
I am using solr version 3.4
The fact that they're python and java is largely beside the point I think.
Solr just sees a URL, the fact that your Python app gets in there
first and "does stuff" with the query wouldn't affect Solr at all.
Also, I tend to like keeping Solr fairly lean so any work I can offload to
the application
Right. See:
http://lucene.apache.org/core/old_versioned_docs/versions/3_4_0/api/contrib-join/org/apache/lucene/search/join/package-summary.html
I guess "refinement" wasn't a good word choice. The basic join stuff
has been in Solr for a while (2272), but 3076 refers to exposing
functionality that
as for version below 4.0, it's not possible because lucene's score
model. position information is stored, but only used to support phrase
query. it just tell us whether a document is matched, but we can boost
a document. The similar problem is : how to implement proximity boost.
for 2 search terms,
By default, the default merge policy (TieredMergePolicy) won't create
the CFS if the segment is "very large" (> 10% of the total index
size). Likely that's what you are seeing?
If you really must have a CFS (how come?) then you can call
TieredMergePolicy.setNOCFSRatio(1.0) -- not sure how/where t
I discounted geonames to start with but it actually looks pretty good. I may
be stretching the limit of my question here, but say I did go with geonames,
if I go back to my model and add a bit:
Search for "London"->Convert "London to Long/Lat"->Send Query to
Solr>->Return Query>
Since
I have heard that GeoNames is a great source for name/location information.
They even have UK postal codes:
http://www.geonames.org/postal-codes/postal-codes-uk.html
-- Jack Krupansky
-Original Message-
From: Michael Della Bitta
Sent: Thursday, May 03, 2012 3:32 PM
To: solr-user@lucen
Hello All,
We are currently looking for a Senior Solr Search Architect to work on a
large scale real time search system.
Our search system scales to petabytes of data, and its a pretty interesting
design problem.
You can get more information on the job and apply using the link, put in my
name as
Hi,
What is the error that you are getting ?
ROWNUM works fine with DIH, I have tried and tested it with Solr 3.1.
One thing that comes to my mind is the query that you are using to
implement ROWNUM.
Do you replaced the "<" in the query by a "<" in dataconfig.xml ?
like "ROMNUM < =100" ?
On Thu,
I am doing database import using solr DIH. I would like to limit the solr
indexing to specific number. In other words If Solr reaches indexing 100
records I want to database import to stop importing.
Not sure if there is any particular setting that would tell solr that I only
want to import 100 ro
I am indexing records from database using DIH. The content of my record is in
html format. When I use browse
I would like to show the content in html format, not in text format. Any
ideas?
--
View this message in context:
http://lucene.472066.n3.nabble.com/how-to-present-html-content-in-browse-tp
this is called geocoding and is properly a subject for GIS types.
it can be non trivial and the data you need to set it up may not be cheap.
i can't address the UK application, but i am somewhat familiar with the US
problem space, and in the US 5 digit postal ("zip") codes don't map to
discreet loc
I discovered the schema.xml file about 2 minutes before I got your response. It
was very enlightening:-)
thanks for the tips about dynamicFields!
On May 3, 2012, at 1:02 PM, Jack Krupansky wrote:
> Those three field names are already in the Solr example schema. Either
> manually add your desi
Thanks a lot to help me to find one solution.
I am going to use multi-valued string field.
regards,
Carlos Andres Garcia Garcia
@grayknight14
> Hi all,
>
>
>
> I need suggetion:
>
>
>
> I
>
> Hi all,
>
>
>
> I need suggetion:
>
>
>
> I have many title like:
>
>
>
> 1 bomb blast in kabul
>
> 2 kabul bomb blast
>
> 3 3 people killed in serial bomb blast in kabul
>
>
>
> I want 2nd result should come first while user search by "kabul".
>
>
Hi,
Can someone officially confirm that it is not supported by current Solr
version
to use both EmbeddedSolrServer(For Full indexing) and
StreamingUpdateSolrServer(For Incremental indexing )
to update the same index?
How can I request for enhancement in the next version?
I think that this requir
Have you tried www.geonames.org ?
- Michael
On Thu, 2012-05-03 at 12:20 -0700, Spadez wrote:
> Hi,
>
> This is quite a challenge. I know there are situations when you can get by
> with google maps api or similar, but they limit the number of requests and I
> need more than that, unfortunatly fo
Hi,
This is quite a challenge. I know there are situations when you can get by
with google maps api or similar, but they limit the number of requests and I
need more than that, unfortunatly for the full service they charge a
fortune!
So, going back to my question, does anyone have any ideas or su
On Thu, May 3, 2012 at 3:01 PM, Tomás Fernández Löbbe
wrote:
> Is this still true? Assuming that I know that there hasn't been updates or
> that I don't care to see a different version of the document, are the term
> QP or the raw QP faster than the real-time get handler?
Sort of different things
Is this still true? Assuming that I know that there hasn't been updates or
that I don't care to see a different version of the document, are the term
QP or the raw QP faster than the real-time get handler?
On Fri, Mar 11, 2011 at 3:12 PM, Yonik Seeley wrote:
> On Fri, Mar 11, 2011 at 5:58 PM, onl
Awesome, thanks!
On Thu, May 3, 2012 at 2:32 PM, Yonik Seeley wrote:
> On Thu, May 3, 2012 at 2:26 PM, okayndc wrote:
> [...]
> > I've experimented with this:
> > {!ex=dt key=Categories and Stuff}category
> >
> > I'm not really sure what 'ex=dt' does but it's obvious that 'key' is the
> > desire
On Thu, May 3, 2012 at 2:26 PM, okayndc wrote:
[...]
> I've experimented with this:
> {!ex=dt key=Categories and Stuff}category
>
> I'm not really sure what 'ex=dt' does but it's obvious that 'key' is the
> desired display name? If there are spaces in the 'key' value, the display
> name gets cut o
Hi,
I am using solr dataimportHandler for doing data import. Since the amount of
data volume is large and changes frequently. I used the suggested approach
described in
http://wiki.apache.org/solr/DataImportHandlerDeltaQueryViaFullImport for delta
import. However I noticed that after the first
Hello,
Is there a way to change the display name (that contains spaces or special
characters) for a facet without changing the value of the facet field? For
example if my facet field name is 'category', I want to change the display
name of the facet to 'Categories and Stuff'
I've experimented wit
Those three field names are already in the Solr example schema. Either
manually add your desired fields to the schema, change their names (column
vs. sourceColName) to fields that do exist in your Solr schema, give them
names that end with one of the dynamicField suffixes (such as "*_s"), or
en
You're on the right track. Solr knows nothing about converting
post codes to lat/lon, you have to do that outside the request
and submit a standard distance query.
Of course this is a bit "interesting". I assume the post codes aren't
perfectly circular (or rectangular for that matter) so you'll ge
I have a data-config.xml declaring some entities and no matter what fields I
declare in the entities the only ones it will index are id, name, and
description. So fields like firstname, email, url don't appear in the index.
They also don't appear in the schema browser. Am I doing something wrong
Erick,
Generally I agree, but could you please expand your definition "is
refinement". What does it mean?
I suggested SOLR-3076, because "index time" has been mention.
On Thu, May 3, 2012 at 5:35 PM, Erick Erickson wrote:
> Solr join has been implemented for quite some time, see:
> https://issu
Hi
Thanks for the information. Steve, the xml is visible in nabble itself, but
that's not a solution for people receiving the mails.
Robert, I tried to implement the factory to deal with german umlaut and
stuff, but am now back with an adapted This should work. According to
http://wiki.apache.o
On Thu, May 3, 2012, at 11:10 AM, Erick Erickson wrote:
> I've never seen lib directives nested, I doubt they're necessary
> and it's vaguely possible that this is not intentionally supported.
>
> I'd try un-nesting them personally.
>
changing to,
still works, doesn't appear to c
On 5/1/2012 6:55 AM, geeky2 wrote:
you said, you don't use autocommit. if so - then why don't you use / like
autocommit?
It's not really that I don't like it, I just don't need it. I think
that it actually caused me problems when I first started using Solr
(1.4.0), but that's been long enou
I've never seen lib directives nested, I doubt they're necessary
and it's vaguely possible that this is not intentionally supported.
I'd try un-nesting them personally.
Best
Erick
On Thu, May 3, 2012 at 10:35 AM, wrote:
>
>
> On Wed, May 2, 2012, at 02:16 PM, Robert Petersen wrote:
>> I don't
Just for a baseline, how much memory is available in the JVM (using jconsole
or something similar) before you do your first query, and then after your
first query (that has these 50-70 facets), and then after a few different
queries (different facets.) Just to see how close you are to "the edge"
On 5/1/2012 8:57 AM, Rahul R wrote:
Hello,
A related question on this topic. How do I programmatically find the total
number of documents across many shards ? For EmbeddedSolrServer, I use the
following command to get the total count :
solrSearcher.getStatistics().get("numDocs")
With distributed
Why not pass the parameters using "?parameter1=value1¶meter2=value2" ?
mvg,
Jasper
On Thu, Apr 26, 2012 at 9:03 PM, Paul Libbrecht wrote:
> Or write your own query component mapping /solr/* in the web.xml, exposing
> the request by a thread-local through a filter, and reading this setting the
Your issue may relate to the migration of the Velocity response writer back
to contrib that occurred in Solr 3.5.
You can read about it here:
http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201202.mbox/%3ccb50294a.2434%25dipti.srivast...@apollogrp.edu%3E
And in the 3.5 release notes:
On Wed, May 2, 2012, at 02:16 PM, Robert Petersen wrote:
> I don't know if this will help but I usually add a dataDir element to
> each cores solrconfig.xml to point at a local data folder for the core
> like this:
after a bit of digging, your suggestion PLUS a change to the 'lib dir'
specificat
ok, not that easy :)
I did not test it myself but it seems that you could use an XSL
preprocessing with the 'xsl' option in your XPathEntityProcessor :
http://wiki.apache.org/solr/DataImportHandler#Configuration_in_data-config.xml-1
You could transform the author part as you wish and then import
If query-side multi-term synonyms are important to your application, your
best bet may be to implement a preprocessor that expands them to an OR
sequences of phrases before submitting the query to Solr. That would also
give you an opportunity to boost a preferred synonym.
For example, a user q
Hi,
I'm finding it a bit hard to get my head around this. Say I am putting items
on a map. This is how I am thinking it would work:
A user submits an "item" and specifies the location as "London"
On submission, I run a process to convert "London" to a Long/Lat which is
stored in the database
How
On Thu, May 3, 2012 at 9:35 AM, OliverS wrote:
> Hello
>
> I read and tried a lot, but somehow I don't fully understand and it doesn't
> work. I'm working on solr 4.0 (latest trunk) and use
> ICUCollationKeyFilterFactory for my main field type. Now, wildcard queries
> don't work, even though ICUCo
Hi Oliver,
Nabble.com stripped out your analysis chain XML before sending your message to
the mailing list. My suggestion: stop using Nabble.
(I've described this problem to their support people a couple of times, and
they apparently just don't care, since it still persists, years later.)
Ahhh, you're right. Shows what happens when I work from memory
Thanks.
Erick
On Wed, May 2, 2012 at 4:26 PM, Jason Rutherglen
wrote:
>> BTW, in 4.0, there's DocumentWriterPerThread that
>> merges in the background
>
> It flushes without pausing, but does not perform merges. Maybe you're
> t
Hello
I read and tried a lot, but somehow I don't fully understand and it doesn't
work. I'm working on solr 4.0 (latest trunk) and use
ICUCollationKeyFilterFactory for my main field type. Now, wildcard queries
don't work, even though ICUCollationKeyFilterFactory seems to be
http://lucene.apache.o
Solr join has been implemented for quite some time, see:
https://issues.apache.org/jira/browse/SOLR-2272
but only on trunk.
3076 is a refinement as I understand it.
FWIW
Erick
On Thu, May 3, 2012 at 3:01 AM, Mikhail Khludnev
wrote:
> Hello,
>
> Here is my favorite ones:
> http://blog.mikemccand
Koji,
Using the way i have specified datadir, i was able to see solr reading my file.
It didnt have any issue reading, but it was not serving the results using
"/elevate".
I looked at the ElevationComponent java code, i didn't see any issue with the
code either.
I need elevation to work with s
Jack,
I am also using synonyms at query side, but so far i am able to use only single
words to work, multi words is not working for me. I didn't want to use synonyms
during indexing, to avoid re indexing.
Is there a way for solr to support multi words?
Ex:
John Cena, John, Cena
Or
Triple H, DX,
Is what I want even possible with XPathEntityProcessor?
It sort of works now - I didn't realize the "flatten" attribute is an attribute
of field instead of entity.
BUT it's still not what I would like.
The XML looks like below and it's nested within
/MedlineCitationSet/MedlineCitation/Article/
hello all,
enviornment: centOS and solr 3.5
i want to make sure i understand the difference between snapshots and solr
replication.
snapshots are "old school" and have been deprecated with solr replication
"new school".
do i have this correct?
btw: i have replication working (now), between my
Hi,
I'm using the following configuration for solr
Solr Specification Version: 1.4.0.2010.01.13.08.09.44
Solr Implementation Version: 1.5-dev exported - yonik - 2010-01-13 08:09:44
Lucene Specification Version: 2.9.1-dev
Lucene Implementation Version: 2.9.1-dev 888785 - 2009-12-09 18:03:31
C
Thank you for the tips :)
Gary
Le 02/05/2012 21:26, Chris Hostetter a écrit :
: String dateString = "20101230";
: SimpleDateFormat sdf = new SimpleDateFormat("MMdd");
: Date date = sdf.parse(dateString);
: doc.addField("date", date);
:
: In the index, the date "20101230" is saved as "2010-1
Hi David,
what do you want to do with the 'commonField' option ?
Is it possible to have the part of the schema for the author field please ?
Is the author field stored ?
Ludovic.
-
Jouve
France.
--
View this message in context:
http://lucene.472066.n3.nabble.com/correct-XPATH-syntax-tp3951
thanks for all of the advice / help.
i appreciate it ;)
--
View this message in context:
http://lucene.472066.n3.nabble.com/should-slave-replication-be-turned-off-on-during-master-clean-and-re-index-tp3945531p3959088.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi.
Compression is a good suggestion. All large dictionaries are compressed well
below 1MB with GZIP. Where should this be implemented? SolrZkClient or
ZkController? Which good compressor is already in Solr's lib? And what's the
difference between SolrZkClient setData and create? Should it auto
Guessing from the message,
java.lang.RuntimeException: [solrconfig.xml]
indexDefaults/mergePolicy: missing mandatory attribute 'class'
somewhere in your solr configs you have something like:
10
10
rather than
10
10
Hi
Ive started using solr 3.6 and would like to use the /browse requestHandler
as i normally do.
But it just gives me some lazyloading error when trying to reach
/solr/browse.
This would normally work in solr 3.4.
So my question is what setup is needed for the velocity responseWriter
That's the standard way, it's actually pretty efficient. Why is this
a concern? Just the verbosity of the getRestults()?
Best
Erick
On Wed, May 2, 2012 at 11:58 AM, vybe3142 wrote:
> I can achieve this by building a query with start and rows = 0, and using
> .getResults().getNumFound().
>
> Are
Hmmm, can we have some more details here? What version of Solr? What
exactly did you do in the UI? What was the state of your index (i.e. adding
documents from some other process? etc.).
Best
Erick
On Wed, May 2, 2012 at 8:17 AM, mechravi25 wrote:
> Hi,
>
>
> When I tried to remove a data from U
On May 3, 2012, at 5:15 AM, Markus Jelsma wrote:
> Hi,
>
> We've increased Zookeepers znode size limit to accomodate for some larger
> dictionaries and other files. It isn't the best idea to increase the maximum
> znode size. Any plans for splitting up larger files and storing them with
> mul
Yes numfound is 0
I tried your way of defining solrdir, didn't work either.
Right now I placed my elevate.xml in conf dir, and wrote a script to reload
solr webapp. We will run the script once a day.
- Original Message -
From: Koji Sekiguchi [mailto:k...@r.email.ne.jp]
Sent: Wednesday
Great, thanks Otis and Erick for your responses
I will take a look at SPM
Thanks
Prabhu
-Original Message-
From: Otis Gospodnetic [mailto:otis_gospodne...@yahoo.com]
Sent: 03 May 2012 00:02
To: solr-user@lucene.apache.org
Subject: Re: Solr Merge during off peak times
Hello Prabhu,
Look
Hi,
I've migrated the search servers to the latest stable release (SOLR-3.5)
from SOLR-1.4.1.
We've fully recreated the index for this. After index completes, when im
optimizing the index then it is not merging the index into a single .cfs
file as was being done with 1.4.1 version.
We've set the
Hi,
We've increased Zookeepers znode size limit to accomodate for some larger
dictionaries and other files. It isn't the best idea to increase the maximum
znode size. Any plans for splitting up larger files and storing them with
multi? Does anyone have another suggestion?
Thanks,
Markus
On Wed, May 2, 2012 at 9:35 PM, Emes, Matthew (US - Irvine)
wrote:
> Hi:
>
> I have been working on an integration project involving Solr 3.5.0 that
> dynamically registers cores as needed at run-time, but does not contain any
> cores by default. The current solr.xml configuration file is:-
>
>
>
Hello,
Here is my favorite ones:
http://blog.mikemccandless.com/2012/01/searching-relational-content-with.html
https://issues.apache.org/jira/browse/SOLR-3076
On Thu, May 3, 2012 at 10:17 AM, tamanjit.bin...@yahoo.co.in <
tamanjit.bin...@yahoo.co.in> wrote:
> Hi,
> I just wanted to get some info
74 matches
Mail list logo