Hi,

With auto-committing disabled we can now index many millions of documents in our test environment on a 5-node cluster with 5 shards and a replication factor of 2. The documents are uploaded from map/reduce. No significant changes were made to solrconfig and there are no update processors enabled. We are using a trunk revision from this weekend.

The indexing speed is well below what we are used to see, we can easily index 5 millions documents on a non-cloud enabled Solr 3.x instance within an hour. What could be going on? There aren't many open TCP connections and the number of file descriptors is stable and I/O is low but CPU-time is high! Each node has two Solr cores both writing to their dedicated disk.

The indexing speed is stable, it was slow at start and still is. It's now running for well over 6 hours and only 3.5 millions documents are indexed. Another strange detail is that the node receiving all incoming documents (we're not yet using a client side Solr server pool) has a much larger disk usage than all other nodes. This is peculiar as we expected all replica's to be a about the same size.

The receiving node has slightly higher CPU than the other nodes but the thread dump shows a very large amount of threads of type cmdDistribExecutor-8-thread-292260 (295090) with 0-100ms CPU-time. At the top of the list these threads all have < 20ms time but near the bottom it rises to just over 100ms. All nodes have a couple of http-80-30 (121994) threads with very high CPU-time each.

Is this a known issue? Did i miss something? Any ideas?

Thanks

Reply via email to