Why can't they be? I've seen indexing documents on the order of 6k/second.
Is the problem with getting the data out of the DB (i.e. how long does the
SQL take to execute) or on the Solr indexing side? If the former, you could
consider writing a SolrJ program that accessed the database and sent the
docs to Solr. If you can write the program such that you could run N copies
(each updating only a portion of the data) you could run them in parallel.

But I suspect that the issue is probably your SQL, and before looking at
Solr solutions I'd try to verify that.

Best,
Erick


On Fri, Nov 23, 2012 at 11:03 AM, zbindigonzales <sandrozbin...@gmail.com>wrote:

> Hi there.
>
> In our company we use Apache Solr 4 to index data from the database via the
> databaseimport handler.
> The data we are indexing is a denormalzied table (Patient, Visit, Study,
> Image).
>
> One requirement is to be near realtime. For that we use softcomits every
> second.
> The index size is about 3M documents.
>
> The problem now is that if a useres modify a entry on a the patient table
> there may be 1000 rows affected (due to denormalization)  wich can't be
> indexed in near realtime.
>
> Did anyone had a similar problem and got a workarround to speed up delta
> imports ?
>
> Best Regards
>
>
>
>
>
>
>
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Solr-Near-Realtime-with-denormalized-Data-tp4022072.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>

Reply via email to