Hello Shawn, I'm sorry to diverge this thread a little bit. But could please point me to resources that explain deeply how this process of OS using the non-java memory to cache index data?
> Whatever RAM is left over after you give 12GB to Java for Solr will be > used automatically by the operating system to cache index data on the > disk. Solr is completely reliant on that caching for good performance. I'm puzzled as to why the physical memory of solr's host machine is always used up and I think some resources on that would help me understand it. Thanks On Tue, Oct 6, 2015 at 5:07 PM, Siddhartha Singh Sandhu < sandhus...@gmail.com> wrote: > Thank you for helping out. > > Further inquiry: I am committing records to my solr implementation and they > are not getting showing up in my search. I am search on the default id. > Is this related to the fact that I dont have enough memory so my SOLR is > taking a lot of time to actually making the indexed documents available > instantly. > > I also looked at the solr log when I sent in my curl commit with my > record(which I can not see in the SOLR instance even after sending it > repeatedly), but it didn't through an error. > > I got this as my response on insertion of that record: > > {"responseHeader":{"status":0,"QTime":57}} > > Thank you. > > Sid. > > On Tue, Oct 6, 2015 at 3:21 PM, Shawn Heisey <apa...@elyograg.org> wrote: > > > On 10/6/2015 8:18 AM, Siddhartha Singh Sandhu wrote: > > > A have a few questions about optimize. Is the search index fully > > searchable > > > after a commit? > > > > If openSearcher is true on the commit, then changes to the index > > (additions, replacements, deletions) will be visible when the commit > > completes. > > > > > How much time does one have to wait in case of a hard commit for the > > index > > > to be available? > > > > This is impossible to answer. It will take as long as it takes, and the > > time will depend on many factors, so it is nearly impossible to > > predict. The only way to know is to try it ... and the number you get > > on one test may be very different than what you actually see once the > > system is in production. > > > > > I have an index of 180G. Do I need to hit the optimize on this chunk. > > This > > > is a single core. Say I cannot get in a cloud env because of cost but > > this > > > is a fairly large > > > amazon machine where I have given SOLR 12G of memory. > > > > Whatever RAM is left over after you give 12GB to Java for Solr will be > > used automatically by the operating system to cache index data on the > > disk. Solr is completely reliant on that caching for good performance. > > A perfectly ideal system for that index and heap size would have 192GB > > of RAM, which is enough to entirely cache the index. I personally > > wouldn't expect good performance with less than 96GB. Some systems with > > a 180GB index and a 12GB heap might be OK with 64GBtotal memory, while > > others with the same size index will require more. > > > > > > > https://lucidworks.com/blog/sizing-hardware-in-the-abstract-why-we-dont-have-a-definitive-answer/ > > > > If the index is on SSD, then RAM is *slightly* less important, and > > performance usually goes up with SSD ... but an SSD cannot completely > > replace RAM, because RAM is much faster. With SSD, you can get away > > with less RAM than you can on a spinning disk system, but depending on a > > bunch of factors, it may not be a LOT less RAM. > > > > https://wiki.apache.org/solr/SolrPerformanceProblems > > > > Optimizing the index is almost never necessary with recent versions. In > > almost all cases optimizing will get you a performance increase, but it > > comes at a huge cost in terms of resource utilization to DO the > > optimize. While the optimize is happening performance will likely be > > worse, possibly a LOT worse. Newer versions of Solr (Lucene) have > > closed the gap on performance with non-optimized indexes, so it doesn't > > gain you as much in performance as it did in earlier versions. > > > > Thanks, > > Shawn > > > > >