I am using 64 bit JVM and we are going out of memory in extraction phase where 
TIKA assigns the content after extracting to SOLRInputDocument in the pipeline 
which gets loaded in memory. 

We are using released 3.1 version of SOLR.

Thanks,
Tirthankar

-----Original Message-----
From: simon [mailto:mtnes...@gmail.com] 
Sent: Tuesday, August 30, 2011 1:23 PM
To: solr-user@lucene.apache.org
Subject: Re: Document Size for Indexing

what issues exactly ?

are you using 32 bit Java ? That will restrict the JVM heap size to 2GB max.

-Simon

On Tue, Aug 30, 2011 at 11:26 AM, Tirthankar Chatterjee < 
tchatter...@commvault.com> wrote:

> Hi,
>
> I have a machine (win 2008R2) with 16GB RAM, I am having issue 
> indexing 1/2GB files. How do we avoid creating a SOLRInputDocument or 
> is there any way to directly use Lucene Index writer classes.
>
> What would be the best approach. We need some suggestions.
> 
> Thanks,
> Tirthankar
>
>
> ******************Legal Disclaimer***************************
> "This communication may contain confidential and privileged material 
> for the sole use of the intended recipient. Any unauthorized review, 
> use or distribution by others is strictly prohibited. If you have 
> received the message in error, please advise the sender by reply email 
> and delete the message. Thank you."
> *********************************************************
******************Legal Disclaimer***************************
"This communication may contain confidential and privileged
material for the sole use of the intended recipient. Any
unauthorized review, use or distribution by others is strictly
prohibited. If you have received the message in error, please
advise the sender by reply email and delete the message. Thank
you."
*********************************************************

Reply via email to