What can I do temporarily in this situation? It seems like I must eventually move to a distributed setup. I am sorting on dynamic float fields.
On Wed, Aug 17, 2011 at 3:01 PM, Yonik Seeley <yo...@lucidimagination.com>wrote: > On Wed, Aug 17, 2011 at 5:56 PM, Jason Toy <jason...@gmail.com> wrote: > > I've only set set minimum memory and have not set maximum memory. I'm > doing > > more investigation and I see that I have 100+ dynamic fields for my > > documents, not the 10 fields I quoted earlier. I also sort against those > > dynamic fields often, I'm reading that this potentially uses a lot of > > memory. Could this be the cause of my problems and if so what options do > I > > have to deal with this? > > Yes, that's most likely the problem. > Sorting on an integer field causes a FieldCache entry with an > int[maxDoc] (i.e. 4 bytes per document in the index, regardless of if > it has a value for that field or not). > Sorting on a string field is 4 bytes per doc in the index (the ords) > plus the memory to store the actual unique string values. > > -Yonik > http://www.lucidimagination.com > > > > > On Wed, Aug 17, 2011 at 2:46 PM, Markus Jelsma > > <markus.jel...@openindex.io>wrote: > > > >> Keep in mind that a commit warms up another searcher and potentially > >> doubling > >> RAM consumption in the back ground due to cache warming queries being > >> executed > >> (newSearcher event). Also, where is your Xmx switch? I don't know how > your > >> JVM > >> will behave if you set Xms > Xmx. > >> > >> 65m docs is quite a lot but it should run fine with 3GB heap allocation. > >> > >> It's a good practice to use a master for indexing without any caches and > >> warm- > >> up queries when you exceed a certain amount of documents, it will bite. > >> > >> > I have a large ec2 instance(7.5 gb ram), it dies every few hours with > out > >> > of heap memory issues. I started upping the min memory required, > >> > currently I use -Xms3072M . > >> > I insert about 50k docs an hour and I currently have about 65 million > >> docs > >> > with about 10 fields each. Is this already too much data for one box? > How > >> > do I know when I've reached the limit of this server? I have no idea > how > >> > to keep control of this issue. Am I just supposed to keep upping the > min > >> > ram used for solr? How do I know what the accurate amount of ram I > should > >> > be using is? Must I keep adding more memory as the index size grows, > I'd > >> > rather the query be a little slower if I can use constant memory and > have > >> > the search read from disk. > >> > > > > > > > > -- > > - sent from my mobile > > 6176064373 > > > -- - sent from my mobile 6176064373