Hi, Gilbert:

Thanks for your tip! I just tried it. Unfortunately, it does not work for me. I still get the OOM exception.

How large was your dataset?  And what were your machine specs?

Cheers,

- Bill

--------------------------------------------------
From: "Gilbert Boyreau" <gboyr...@andevsol.com>
Sent: Tuesday, October 27, 2009 11:54 AM
To: <solr-user@lucene.apache.org>
Subject: Re: DIH out of memory exception

Hi,

I got the same problem using DIH with a large dataset in MySql database.

Following : http://dev.mysql.com/doc/refman/5.1/en/connector-j-reference-implementation-notes.html, and looking at the java code, it appears that DIH use PreparedStatement in the JdbcDataSource.

I set the batchsize parameter to -1 and it solved my problem.

Regards.
Gilbert.

William Pierce a écrit :
Folks:

My db contains approx 6M records -- on average each is approx 1K bytes. When I use the DIH, I reliably get an OOM exception. The machine has 4 GB ram, my tomcat is set to use max heap of 2G. The option of increasing memory is not tenable coz as the number of documents grows I will be back in this situation. Is there a way to batch the documents? I tried setting the batchsize parameter to 500 on the <dataSource> tag where I specify the jdbc parameters. This had no effect.

Best,

- Bill



Reply via email to