2GB is a rather small heap. Our production systems run with 8GB and smaller indexes than that. Our dev and test systems run with 6GB heaps.
wunder On Jun 10, 2013, at 9:52 PM, gururaj kosuru wrote: > Hello, > I have recently started using solr 3.4 and have a standalone > system deployed that has 40,000,000 data rows with 3 indexed fields > totalling around 9 GB. I have given a Heap size of 2GB and I run the > instance on Tomcat on an i7 system with 4 GB RAM. My queries involve > searching among the indexed fields using the keyword 'OR' to search for > multiple words in a single field and the keyword 'AND' to get the > intersection of multiple matches from different fields. The problem I face > is an out of memory exception that happens when the solr core has been > queried for a long time and I am forced to restart the solr instance. My > primary questions are : > > 1. Can I make any changes in my query as I have noticed that if I divide my > query into parts eg- if my query has "A and B and C", executing only A > gives out 75,000 results. However, If I run the whole query, I get an out > of memory error in SegmentNorms.java at line 156 which allocates a new byte > array of size count. > > 2. Does my index need to be able to fit into the RAM in one go? > > 3. Will moving to Solr Cloud solve the out of memory exception issue and if > so, what will be the ratio of my RAM size to the shard size that should be > used? > > Thanks a lot, > Gururaj -- Walter Underwood wun...@wunderwood.org