Ideally you would want to use SOLRJ or other interface which can catch
exceptions/error and re-try them.


On Tue, Feb 26, 2013 at 3:45 PM, Walter Underwood <wun...@wunderwood.org>wrote:

> I've done exactly the same thing. On error, set the batch size to one and
> try again.
>
> wunder
>
> On Feb 26, 2013, at 12:27 PM, Timothy Potter wrote:
>
> > Here's what I do to work-around failures when processing batches of
> updates:
> >
> > On client side, catch the exception that the batch failed. In the
> > exception handler, switch to one-by-one mode for the failed batch
> > only.
> >
> > This allows you to isolate the *bad* documents as well as getting the
> > *good* documents in the batch indexed in Solr.
> >
> > This assumes most batches work so you only pay the one-by-one penalty
> > for the occasional batch with a bad doc.
> >
> > Tim
> >
> > On Tue, Feb 26, 2013 at 12:08 PM, Isaac Hebsh <isaac.he...@gmail.com>
> wrote:
> >> Hi.
> >>
> >> I add documents to Solr by POSTing them to UpdateHandler, as bulks of
> <add>
> >> commands (DIH is not used).
> >>
> >> If one document contains any invalid data (e.g. string data into numeric
> >> field), Solr returns HTTP 400 Bad Request, and the whole bulk is failed.
> >>
> >> I'm searching for a way to tell Solr to accept the rest of the
> documents...
> >> (I'll use RealTimeGet to determine which documents were added).
> >>
> >> If there is no standard way for doing it, maybe it can be implemented by
> >> spiltting the <add> commands into seperate HTTP POSTs. Because of using
> >> auto-soft-commit, can I say that it is almost equivalent? What is the
> >> performance penalty of 100 POST requests (of 1 document each) againt 1
> >> request of 100 docs, if a soft commit is eventually done.
> >>
> >> Thanks in advance...
>
> --
> Walter Underwood
> wun...@wunderwood.org
>
>
>
>


-- 
Anirudha P. Jadhav

Reply via email to