Oops. My wording was poor. My reference to those who don't research the matter was pointing at a large number of engineers I have worked with; not this list.
-----Original Message----- From: Will Martin [mailto:wmartin...@gmail.com] Sent: Wednesday, October 29, 2014 6:38 PM To: 'solr-user@lucene.apache.org' Subject: RE: Solr Memory Usage This command only touches OS level caches that hold pages destined for (or not) the swap cache. Its use means that disk will be hit on future requests, but in many instances the pages were headed for ejection anyway. It does not have anything whatsoever to do with Solr caches. It also is not fragmentation related; it is a result of the kernel managing virtual pages in an "as designed manner". The proper command is #sync; echo 3 >/proc/sys/vm/drop_caches. http://linux.die.net/man/5/proc I have encountered resistance on the use of this on long-running processes for years ... from people who don't even research the matter. -----Original Message----- From: Toke Eskildsen [mailto:t...@statsbiblioteket.dk] Sent: Wednesday, October 29, 2014 3:06 PM To: solr-user@lucene.apache.org Subject: RE: Solr Memory Usage Vijay Kokatnur [kokatnur.vi...@gmail.com] wrote: > For the Solr Cloud setup, we are running a cron job with following > command to clear out the inactive memory. It is working as expected. > Even though the index size of Cloud is 146GB, the used memory is always below 55GB. > Our response times are better and no errors/exceptions are thrown. > (This command causes issue in 2 Shard setup) > echo 3 > /proc/sys/vm/drop_caches As Shawn points out, this is under normal circumstances a very bad idea, but... > Has anyone faced this issue before? We did have some problems on a 256GB machine churning terabytes of data through 40 concurrent Tika processes and into Solr. After some days, performance got really bad. When we did a top, we noticed that most of the time was used in the kernel (the 'sy' on the '%Cpu(s):'-line). The drop_caches trick worked for us too. Our systems guys explained that it was because of virtual memory space fragmentation, so the OS had to spend a lot of resources just bookkeeping memory. Try keeping an eye on the fraction of processing power spend on the kernel from you clear the cache until it performance gets bad again. If it rises drastically, you might have the same problem. - Toke Eskildsen