Hey Folks, we today hit the same error three times, a REPLACENODE call was not successful.
Here is our scenario: 3 Node Solrcloud cluster running in Kubernetes on top of AWS. Today we wanted to rotate the underlying storage (increased from 50gb to 300gb). After we rotated one node we tried to replace with this call: • curl 'solr-2.solr-discovery.default.svc.cluster.local:8983/solr/admin/collections?action=REPLACENODE&source=solr-2.solr-discovery.default.svc.cluster.local.:8983_solr&target=solr-2.solr-discovery.default.svc.cluster.local.:8983_solr&async=4495d85b-0aa4-45ab-8067-9d7d4da375d3' • curl 'solr-2.solr-discovery.default.svc.cluster.local:8983/solr/admin/collections?action=REQUESTSTATUS&requestid=4495d85b-0aa4-45ab-8067-9d7d4da375d3’ The error we got was: <?xml version="1.0" encoding="UTF-8"?> <response> <lst name="responseHeader"><int name="status">0</int><int name="QTime">28</int></lst><str name="Operation replacenode caused exception:">java.util.concurrent.RejectedExecutionException:java.util.concurrent.RejectedExecutionException: Task org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$$Lambda$15/509076276@5c9136c8 rejected from org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@1cce4506[Running, pool size = 10, active threads = 10, queued tasks = 0, completed tasks = 0]</str><lst name="exception"><str name="msg">Task org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$$Lambda$15/509076276@5c9136c8 rejected from org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor@1cce4506[Running, pool size = 10, active threads = 10, queued tasks = 0, completed tasks = 0]</str><int name="rspCode">-1</int></lst><lst name="status"><str name="state">failed</str><str name="msg">found [4495d85b-0aa4-45ab-8067-9d7d4da375d3] in failed tasks</str></lst> </response> The problem was that afterwards we had the same shard on the same node twice. One recovered and we had to delete the other one manually. For some collections the REPLACENODE went through and everything was fine again. Can you advice what we did wrong here or which configuration we need to adapt? Thanks Björn