1- each document around 50 kb - 150 kb  (web document)
2-final index is 40 gig
3-jre memory carefully given.



On Mon, Sep 26, 2011 at 9:57 PM, Jaeger, Jay - DOT <jay.jae...@dot.wi.gov>wrote:

> 500 / second would be 1,800,000 per hour (much more than 500K documents).
>
> 1)  how big is each document?
> 2)  how big are your index files?
> 3)  as others have recently written, make sure you don't give your JRE so
> much memory that your OS is starved for memory to use for file system cache.
>
> JRJ
>
> -----Original Message-----
> From: Lord Khan Han [mailto:khanuniver...@gmail.com]
> Sent: Monday, September 26, 2011 6:09 AM
> To: solr-user@lucene.apache.org
> Subject: SOLR Index Speed
>
> Hi,
>
> We have 500K web document and usind solr (trunk) to index it. We have
> special anaylizer which little bit heavy cpu .
> Our machine config:
>
> 32 x cpu
> 32 gig ram
> SAS HD
>
> We are sending document with 16 reduce client (from hadoop) to the stand
> alone solr server. the problem is we couldnt get speedier than the 500 doc
> /
> per sec. 500K document tooks 7-8 hours to index :(
>
> While indexin the the solr server cpu load is around : 5-6  (32 max)  it
> means  %20 of the cpu total power. We have plenty ram ...
>
> I turned of auto commit  and give 8198 rambuffer .. there is no io wait ..
>
> How can I make it faster ?
>
> PS: solr streamindex  is not option because we need to submit javabin...
>
> thanks..
>

Reply via email to