Thanks Otis.

I did increase the number of file descriptors to 22K, but I still get
this problem. I've noticed following so far,

1) As soon as I get to around 1140 index segments (this is total over
multiple cores) I start seeing this problem.
2) When the problem starts occassionally the index request
(solrserver.commit) also fails with the following error,
      java.net.SocketException: Connection reset
3) Whenever the commit fails, I'm able to access Solr by the browser
(http://ets11.co.com/solr). If the commit is succssfull and going on I
get blank page on Firefox. Even the telnet to 8080 fails with
"Connection closed by foreign host."

It does seem like there is some resource issue as it happens only once
we reach a breaking point (too many index segment files) - lsof at
this point usually shows at 1400, but my ulimit is much higher than
that.

I already use compound format for index files. I can also run optimize
occassionally (though not preferred as it blocks the whole index cycle
for a long time). I do want to find out what resource limitation is
causing this and it has to do something with when Indexer is
committing the records where there are large number of segment files.

Any other ideas?

Thanks,
-vivek

On Wed, Apr 15, 2009 at 3:10 PM, Otis Gospodnetic
<otis_gospodne...@yahoo.com> wrote:
>
> One more thing.  I don't think this was mentioned, but you can:
> - optimize your indices
> - use compound index format
>
> That will lower the number of open file handles.
>
>  Otis
> --
> Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
>
>
>
> ----- Original Message ----
>> From: vivek sar <vivex...@gmail.com>
>> To: solr-user@lucene.apache.org
>> Sent: Friday, April 10, 2009 5:59:37 PM
>> Subject: Re: Question on StreamingUpdateSolrServer
>>
>> I also noticed that the Solr app has over 6000 file handles open -
>>
>>     "lsof | grep solr | wc -l"   - shows 6455
>>
>> I've 10 cores (using multi-core) managed by the same Solr instance. As
>> soon as start up the Tomcat the open file count goes up to 6400.  Few
>> questions,
>>
>> 1) Why is Solr holding on to all the segments from all the cores - is
>> it because of auto-warmer?
>> 2) How can I reduce the open file count?
>> 3) Is there a way to stop the auto-warmer?
>> 4) Could this be related to "Tomcat returning blank page for every request"?
>>
>> Any ideas?
>>
>> Thanks,
>> -vivek
>>
>> On Fri, Apr 10, 2009 at 1:48 PM, vivek sar wrote:
>> > Hi,
>> >
>> >  I was using CommonsHttpSolrServer for indexing, but having two
>> > threads writing (10K batches) at the same time was throwing,
>> >
>> >  "ProtocolException: Unbuffered entity enclosing request can not be 
>> > repeated.
>> "
>> >
>> > I switched to StreamingUpdateSolrServer (using addBeans) and I don't
>> > see the problem anymore. The speed is very fast - getting around
>> > 25k/sec (single thread), but I'm facing another problem. When the
>> > indexer using StreamingUpdateSolrServer is running I'm not able to
>> > send any url request from browser to Solr web app. I just get blank
>> > page. I can't even get to the admin interface. I'm also not able to
>> > shutdown the Tomcat running the Solr webapp when the Indexer is
>> > running. I've to first stop the Indexer app and then stop the Tomcat.
>> > I don't have this problem when using CommonsHttpSolrServer.
>> >
>> > Here is how I'm creating it,
>> >
>> > server = new StreamingUpdateSolrServer(url, 1000,3);
>> >
>> > I simply call server.addBeans(...) on it. Is there anything else I
>> > need to do to make use of StreamingUpdateSolrServer? Why does Tomcat
>> > become unresponsive  when Indexer using StreamingUpdateSolrServer is
>> > running (though, indexing happens fine)?
>> >
>> > Thanks,
>> > -vivek
>> >
>
>

Reply via email to