With last night's build of solr, I am trying to use the MailEntityProcessor
to index an email account. However, when I call my dataimport url, I
receive a class cast exception:
INFO: [] webapp=/solr path=/dataimport params={command=full-import} status=0
QTime=44
Jun 16, 2010 8:16:03 PM org.apache
Hey guys,
I'm using Solr 1.4.1 and I've been having some problems lately with code
that adds documents through a CommonsHttpSolrServer. It seems that randomly
the call to theserver.add() will hang. I am currently running my code in a
single thread, but I noticed this would happen in multi threade
Lucene index in
the background)?
Thanks.
On Mon, Jul 12, 2010 at 12:06 PM, Robert Petersen wrote:
> Maybe solr is busy doing a commit or optimize?
>
> -Original Message-
> From: Max Lynch [mailto:ihas...@gmail.com]
> Sent: Monday, July 12, 2010 9:59 AM
> To: solr-user
angs
>
> You could try a master slave setup using replication perhaps, then the
> slave serves searches and indexing commits on the master won't hang up
> searches at least...
>
> Here is the description: http://wiki.apache.org/solr/SolrReplication
>
>
> -Origin
never
opens an IndexWriter.
Thanks!
On Tue, Jul 13, 2010 at 10:52 AM, Max Lynch wrote:
> Great, thanks!
>
>
> On Tue, Jul 13, 2010 at 2:55 AM, Fornoville, Tom > wrote:
>
>> If you're only adding documents you can also have a go with
>> StreamingUpdateSo
I would like to be search against my index, and then *know* which of a set
of given terms were found in each document.
For example, let's say I want to show articles with the word "pizza" or
"cake" in them, but would like to be able to say which of those two was
found. I might use this to handle
ybacking on the highlighter is an OK approach.
>
> If you need it on more docs than that, then probably you should
> customize how your queries are scored to also tally up which docs had
> which terms.
>
> Mike
>
> On Wed, Jul 28, 2010 at 6:53 PM, Max Lynch wrote:
> > I
Is it possible to duplicate a core? I want to have one core contain only
documents within a certain date range (ex: 3 days old), and one core with
all documents that have ever been in the first core. The small core is then
replicated to other servers which do "real-time" processing on it, but the
What I'm doing now is just adding the documents to the other core each night
and deleting old documents from the other core when I'm finished. Is there
a better way?
On Tue, Aug 3, 2010 at 4:38 PM, Max Lynch wrote:
> Is it possible to duplicate a core? I want to have one core
Hi,
I am trying to delete all documents that have null values for a certain
field. To that effect I can see all of the documents I want to delete by
doing this query:
-date_added_solr:[* TO *]
This returns about 32,000 documents.
However, when I try to put that into a curl call, no documents get
now?
>
> -date_added_solr:[* TO *]'
>
> - Original Message -
> From: "Max Lynch"
> To:
> Sent: Thursday, August 26, 2010 6:12 AM
> Subject: Delete by query issue
>
>
> > Hi,
> > I am trying to delete all documents that have null values for
Right now I am doing some processing on my Solr index using Lucene Java.
Basically, I loop through the index in Java and do some extra processing of
each document (processing that is too intensive to do during indexing).
However, when I try to update the document in solr with new fields (using
Sol
means "select everything AND only these documents without a value
> in the field".
>
> On Wed, Aug 25, 2010 at 7:55 PM, Max Lynch wrote:
> > I was trying to filter out all documents that HAVE that field. I was
> trying
> > to delete any documents where that field
from the
CoreContainer:
[java] Aug 25, 2010 10:25:23 PM org.apache.solr.update.SolrIndexWriter
finalize
[java] SEVERE: SolrIndexWriter was not closed prior to finalize(),
indicates a bug -- POSSIBLE RESOURCE LEAK!!!
I'm not sure what exactly triggers that but it's a result of
Hi,
I have a master solr server and two slaves. On each of the slaves I have
programs running that read the slave index, do some processing on each
document, add a few new fields, and commit the changes back to the master.
The problem I'm running into right now is one slave will update one docume
o the master.
>
> Lance
>
> On Sun, Aug 29, 2010 at 6:19 PM, Max Lynch wrote:
> > Hi,
> > I have a master solr server and two slaves. On each of the slaves I have
> > programs running that read the slave index, do some processing on each
> > document, add a few ne
Is there a tokenizer that will allow me to search for parts of a URL? For
example, the search "google" would match on the data "
http://mail.google.com/dlkjadf";
This tokenizer factory doesn't seem to be sufficient:
Hi,
I asked this question a month ago on lucene-user and was referred here.
I have content being analyzed in Solr using these tokenizers and filters:
Basically I want to be able to search against
I have made progress on this by writing my own Analyzer. I basically added
the TokenFilters that are under each of the solr factory classes. I had to
copy and paste the WordDelimiterFilter because, of course, it was package
protected.
On Mon, Oct 4, 2010 at 3:05 PM, Max Lynch wrote:
>
, "1");
> args.put("stemEnglishPossessive", "0");
> args.put("language", "English");
>
> wordDelimiter = new WordDelimiterFilterFactory();
> wordDelimiter.init(args);
> stream = wordDelimiter.create(stream);
>
> --
> Kind
20 matches
Mail list logo