On 22 February 2013 21:32, jimtronic <jimtro...@gmail.com> wrote:
> Yes, these are good points. I'm using solr to leverage user preference data
> and I need that data available real time. SQL just can't do the kind of
> things I'm able to do in solr, so I have to wait until the write (a user
> action, a user preference, etc) gets to solr from the db anyway.

?
Does that not depend on how your data pipeline works?

> I'm kind of curious about how many single documents i can send through via
> the json update in a day. Millions would be nice, but I wonder what the
> upper limit would be.

The devil is in the details, of course, i.e., what is the average size of the
documents, where are they pulled from,what transformations are applied,
what hardware does one have to throw at the problem, etc., but in most
cases millions a day is quite easy to achieve. Other people have posted
benchmarks in this regard, but even with Solr 1.4 we were doing a complete
indexing of ~8 million heavy records (total index size of ~80GB) in 6-8 hours,
and the bottle-neck for us was the database.

Regards,
Gora

Reply via email to