After testing the update?commit=true i now face an error: "Maximum lock
count exceeded". strange this is the first time i see this in the lockfiles
and when doing commit=true
ava.lang.Error: Maximum lock count exceeded
    at
java.base/java.util.concurrent.locks.ReentrantReadWriteLock$Sync.fullTryAcquireShared(ReentrantReadWriteLock.java:535)
    at
java.base/java.util.concurrent.locks.ReentrantReadWriteLock$Sync.tryAcquireShared(ReentrantReadWriteLock.java:494)
    at
java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1368)
    at
java.base/java.util.concurrent.locks.ReentrantReadWriteLock$ReadLock.tryLock(ReentrantReadWriteLock.java:882)
    at
org.apache.solr.update.DefaultSolrCoreState.lock(DefaultSolrCoreState.java:179)
    at
org.apache.solr.update.DefaultSolrCoreState.getIndexWriter(DefaultSolrCoreState.java:124)
    at
org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2.java:658)
    at
org.apache.solr.update.processor.RunUpdateProcessor.processCommit(RunUpdateProcessorFactory.java:102)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalCommit(DistributedUpdateProcessor.java:1079)
    at
org.apache.solr.update.processor.DistributedZkUpdateProcessor.processCommit(DistributedZkUpdateProcessor.java:220)
    at
org.apache.solr.update.processor.LogUpdateProcessorFactory$LogUpdateProcessor.processCommit(LogUpdateProcessorFactory.java:160)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:68)
    at
org.apache.solr.handler.RequestHandlerUtils.handleCommit(RequestHandlerUtils.java:69)
    at
org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:62)
    at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:211)
    at org.apache.solr.core.SolrCore.execute(SolrCore.java:2596)
    at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:799)
    at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:578)
    at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:419)
    at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:351)
    at
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1602)
    at
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:540)
    at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:146)
    at
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548)
    at
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132)
    at
org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:257)
    at
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:1711)
    at
org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:255)
    at
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1347)
    at
org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:203)
    at
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:480)
    at
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:1678)
    at
org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:201)
    at
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1249)
    at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:144)
    at
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:220)
    at
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:152)
    at
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132)
    at
org.eclipse.jetty.rewrite.handler.RewriteHandler.handle(RewriteHandler.java:335)
    at
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132)
    at org.eclipse.jetty.server.Server.handle(Server.java:505)
    at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:370)
    at
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:267)
    at
org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:305)
    at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:103)
    at
org.eclipse.jetty.io.ssl.SslConnection$DecryptedEndPoint.onFillable(SslConnection.java:427)
    at
org.eclipse.jetty.io.ssl.SslConnection.onFillable(SslConnection.java:321)
    at
org.eclipse.jetty.io.ssl.SslConnection$2.succeeded(SslConnection.java:159)
    at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:103)
    at org.eclipse.jetty.io.ChannelEndPoint$2.run(ChannelEndPoint.java:117)
    at
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:333)
    at
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:310)
    at
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:168)
    at
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:126)
    at
org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:366)
    at
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:781)
    at
org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:917)
    at java.base/java.lang.Thread.run(Thread.java:834)

On Tue, Jan 21, 2020 at 10:51 PM Jörn Franke <jornfra...@gmail.com> wrote:

> The only weird thing is I see that for instance I have
> <maxTime>${solr.autoCommit.maxTime:15000}</maxTime>  and similar entries.
> It looks like a template gone wrong, but this was not caused due to an
> internal development. It must have been come from a Solr version.
>
> On Tue, Jan 21, 2020 at 10:49 PM Jörn Franke <jornfra...@gmail.com> wrote:
>
>> It is btw. a Linux system and autosoftcommit is set to -1. However,
>> indeed openSearcher is set to false. A commit is set to true after doing
>> all the updates, but the index is not shrinking. The files are not
>> disappearing during shutdown, but they disappear after starting up again.
>>
>> On Tue, Jan 21, 2020 at 4:04 PM Jörn Franke <jornfra...@gmail.com> wrote:
>>
>>> thanks for the answer I will look into it - it is a possible
>>> explanation.
>>>
>>> > Am 20.01.2020 um 14:30 schrieb Erick Erickson <erickerick...@gmail.com
>>> >:
>>> >
>>> > Jörn:
>>> >
>>> > The only thing I can think of that _might_ cause this (I’m not all
>>> that familiar with the code) is if your solrconfig settings never open a
>>> searcher. Either you need to be sure openSearcher is set to true in the
>>> autocommit section in solrconfig.xml or your autoSoftCommit is set to
>>> something other than -1. Real Time Get requires access to all segments and
>>> it takes a new searcher being opened to release them. Actually, a very
>>> quick test would be to submit 
>>> “http://host:port/solr/collection/update?commit=true”
>>> and see if the index shrinks as a result. You don’t need to change
>>> solrconfig.xml for that test.
>>> >
>>> > If you are opening a new searcher, this is very concerning. There
>>> shouldn’t be anything else you have to set to prevent the index from
>>> growing. Could you check one thing? Compare the directory listing of the
>>> data/index directory just before you shut down Solr and then just after.
>>> What I’m  interested in is whether some subset of files disappears when you
>>> shut down Solr. This assumes you’re running on a *nix system, if Windows
>>> you may have to start Solr again to see the difference.
>>> >
>>> > So if you open a searcher and still see the problem, I can try to
>>> reproduce it. Can you share your solrconfig file or at least the autocommit
>>> and cache portions?
>>> >
>>> > Best,
>>> > Erick
>>> >
>>> >> On Jan 20, 2020, at 5:40 AM, Jörn Franke <jornfra...@gmail.com>
>>> wrote:
>>> >>
>>> >> From what is see it basically duplicates the index files, but does
>>> not delete the old ones.
>>> >> It uses caffeine cache.
>>> >>
>>> >> What I observe is that there is an exception when shutting down for
>>> the collection that is updated - timeout waiting for all directory ref
>>> counts to be released - gave up waiting on CacheDir.
>>> >>
>>> >>>> Am 20.01.2020 um 11:26 schrieb Jörn Franke <jornfra...@gmail.com>:
>>> >>>
>>> >>> Sorry I missed a line - not tlog is growing but the /data/index
>>> folder is growing - until restart when it seems to be purged.
>>> >>>
>>> >>>> Am 20.01.2020 um 10:47 schrieb Jörn Franke <jornfra...@gmail.com>:
>>> >>>>
>>> >>>> Hi,
>>> >>>>
>>> >>>> I have a test system here with Solr 8.4 (but this is also
>>> reproducible in older Solr versions), which has an index which is growing
>>> and growing - until the SolrCloud instance is restarted - then it is
>>> reduced tot the expected normal size.
>>> >>>> The collection is configured to do auto commit after 15000 ms. I
>>> expect the index grows comes due to the usage of atomic updates, but I
>>> would expect that due to the auto commit this does not grow all the time.
>>> >>>> After the atomic updates a commit is done in any case.
>>> >>>>
>>> >>>> I don’t see any error message in the log files, but the growth is
>>> quiet significant and frequent restarts are not a solution of course.
>>> >>>>
>>> >>>> Maybe I am overlooking here a tiny configuration issue?
>>> >>>>
>>> >>>> Thank you.
>>> >>>>
>>> >>>>
>>> >>>> Best regards
>>> >
>>>
>>

Reply via email to