Hi Jan, We are using 64 bit Java, version 1.8.0_191. We started Solr with 6 GB heap size.
Besides Solr, we have ZooKeeper, IIS, Google Chrome and NotePad++ running on the machine. There is still 22 GB of memory left on the server, out of the 32 GB available on the machine. Regards, Edwin On Fri, 25 Jan 2019 at 21:04, Jan Høydahl <jan....@cominvent.com> wrote: > Which java version? 32 or 64 bit? You start Solr with default 512Mb heap > size? > Other software running on the machine? > > -- > Jan Høydahl, search solution architect > Cominvent AS - www.cominvent.com > > > 25. jan. 2019 kl. 13:05 skrev Zheng Lin Edwin Yeo <edwinye...@gmail.com > >: > > > > Hi Jan and Shawn, > > > > For your info, this is another debug query. > > > > "debug":{ > > > > "rawquerystring":"johnny", > > > > "querystring":"johnny", > > > > "parsedquery":"searchFields_tcs:johnny", > > > > "parsedquery_toString":"searchFields_tcs:johnny", > > > > "explain":{ > > > > "192280":"\n12.8497505 = weight(searchFields_tcs:johnny in > > 75730) [SchemaSimilarity], result of:\n 12.8497505 = > > score(doc=75730,freq=4.0 = termFreq=4.0\n), product of:\n 7.5108404 > > = idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq + > > 0.5)) from:\n 473.0 = docFreq\n 865438.0 = docCount\n > > 1.7108272 = tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 - > > b + b * fieldLength / avgFieldLength)) from:\n 4.0 = > > termFreq=4.0\n 1.2 = parameter k1\n 0.75 = parameter b\n > > 26.66791 = avgFieldLength\n 25.0 = fieldLength\n”, > > > > "QParser":"LuceneQParser", > > > > "timing":{ > > > > "time":350.0, > > > > "prepare":{ > > > > "time":0.0, > > > > "query":{ > > > > "time":0.0}, > > > > "facet":{ > > > > "time":0.0}, > > > > "facet_module":{ > > > > "time":0.0}, > > > > "mlt":{ > > > > "time":0.0}, > > > > "highlight":{ > > > > "time":0.0}, > > > > "stats":{ > > > > "time":0.0}, > > > > "expand":{ > > > > "time":0.0}, > > > > "terms":{ > > > > "time":0.0}, > > > > "debug":{ > > > > "time":0.0}}, > > > > "process":{ > > > > "time":348.0, > > > > "query":{ > > > > "time":287.0}, > > > > "facet":{ > > > > "time":0.0}, > > > > "facet_module":{ > > > > "time":0.0}, > > > > "mlt":{ > > > > "time":0.0}, > > > > "highlight":{ > > > > "time":54.0}, > > > > "stats":{ > > > > "time":0.0}, > > > > "expand":{ > > > > "time":0.0}, > > > > "terms":{ > > > > "time":0.0}, > > > > "debug":{ > > > > "time":6.0}}, > > > > "loadFieldValues":{ > > > > "time":0.0}}}} > > > > > > Regards, > > Edwin > > > > > > On Fri, 25 Jan 2019 at 19:52, Zheng Lin Edwin Yeo <edwinye...@gmail.com> > > wrote: > > > >> Hi Jan and Shawn, > >> > >> Please focus on the strange issue that I have described above in more > >> details, summary is as follows: > >> > >> 1. Index customers data, then queries from highlight, select, and all > >> handlers are very fast (less than 50ms) > >> > >> 2. Now index policies data, then queries on polices are very fast BUT > >> queries on customers becomes slow > >> > >> 3. If I reindex customers data, then again queries for customers are > very > >> fast BUT queries on policies becomes slow. > >> > >> How can you explain this behavior? > >> > >> We have never experienced such a strange behavior before Solr 7. > >> > >> Regards, > >> Edwin > >> > >> On Fri, 25 Jan 2019 at 17:06, Zheng Lin Edwin Yeo <edwinye...@gmail.com > > > >> wrote: > >> > >>> Hi Jan, > >>> > >>> Referring to what you have mentioned that the highlighting takes up > most > >>> of the time in the first query from the policies collection, the > >>> highlighting was very fast (less than 50ms) from the time it was > indexed, > >>> till the time after customers collection gets indexed, in which it > slowed > >>> down tremendously. > >>> > >>> Also, the slow down does not just affect on the highlight > requestHandler. > >>> It also affects other requestHandler like select and clustering as > well. > >>> All of them gets the QTime of more than 500ms. This is also proven in > the > >>> latest debug query that we have sent earlier, in which we have set > hl=false > >>> and fl=null. > >>> > >>> Any idea or explanation on this strange behavior? > >>> Thank you for your support, as we look forward to shed some lights on > >>> this issue and to resolve it. > >>> > >>> Regards, > >>> Edwin > >>> > >>> On Thu, 24 Jan 2019 at 23:35, Zheng Lin Edwin Yeo < > edwinye...@gmail.com> > >>> wrote: > >>> > >>>> Hi Jan, > >>>> > >>>> Thanks for your reply. > >>>> > >>>> However, we are still getting a slow QTime of 517ms even after we set > >>>> hl=false&fl=null. > >>>> > >>>> Below is the debug query: > >>>> > >>>> "debug":{ > >>>> "rawquerystring":"cherry", > >>>> "querystring":"cherry", > >>>> "parsedquery":"searchFields_tcs:cherry", > >>>> "parsedquery_toString":"searchFields_tcs:cherry", > >>>> "explain":{ > >>>> "46226513":"\n14.227914 = weight(searchFields_tcs:cherry in > 5747763) [SchemaSimilarity], result of:\n 14.227914 = > score(doc=5747763,freq=3.0 = termFreq=3.0\n), product of:\n 9.614556 = > idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq + 0.5)) > from:\n 400.0 = docFreq\n 6000000.0 = docCount\n 1.4798305 = > tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 - b + b * > fieldLength / avgFieldLength)) from:\n 3.0 = termFreq=3.0\n 1.2 = > parameter k1\n 0.75 = parameter b\n 19.397041 = avgFieldLength\n > 25.0 = fieldLength\n", > >>>> "54088731":"\n13.937909 = weight(searchFields_tcs:cherry in > 4840794) [SchemaSimilarity], result of:\n 13.937909 = > score(doc=4840794,freq=3.0 = termFreq=3.0\n), product of:\n 9.614556 = > idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq + 0.5)) > from:\n 400.0 = docFreq\n 6000000.0 = docCount\n 1.4496675 = > tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 - b + b * > fieldLength / avgFieldLength)) from:\n 3.0 = termFreq=3.0\n 1.2 = > parameter k1\n 0.75 = parameter b\n 19.397041 = avgFieldLength\n > 27.0 = fieldLength\n", > >>>> "QParser":"LuceneQParser", > >>>> "timing":{ > >>>> "time":517.0, > >>>> "prepare":{ > >>>> "time":0.0, > >>>> "query":{ > >>>> "time":0.0}, > >>>> "facet":{ > >>>> "time":0.0}, > >>>> "facet_module":{ > >>>> "time":0.0}, > >>>> "mlt":{ > >>>> "time":0.0}, > >>>> "highlight":{ > >>>> "time":0.0}, > >>>> "stats":{ > >>>> "time":0.0}, > >>>> "expand":{ > >>>> "time":0.0}, > >>>> "terms":{ > >>>> "time":0.0}, > >>>> "debug":{ > >>>> "time":0.0}}, > >>>> "process":{ > >>>> "time":516.0, > >>>> "query":{ > >>>> "time":15.0}, > >>>> "facet":{ > >>>> "time":0.0}, > >>>> "facet_module":{ > >>>> "time":0.0}, > >>>> "mlt":{ > >>>> "time":0.0}, > >>>> "highlight":{ > >>>> "time":0.0}, > >>>> "stats":{ > >>>> "time":0.0}, > >>>> "expand":{ > >>>> "time":0.0}, > >>>> "terms":{ > >>>> "time":0.0}, > >>>> "debug":{ > >>>> "time":500.0}}}}} > >>>> > >>>> Regards, > >>>> Edwin > >>>> > >>>> > >>>> On Thu, 24 Jan 2019 at 22:43, Jan Høydahl <jan....@cominvent.com> > wrote: > >>>> > >>>>> Looks like highlighting takes most of the time on the first query > >>>>> (680ms). You config seems to ask for a lot of highlighting here, > like 100 > >>>>> snippets of max 100000 characters etc. > >>>>> Sounds to me that this might be a highlighting configuration problem. > >>>>> Try to disable highlighting (hl=false) and see if you get back your > speed. > >>>>> Also, I see fl=* in your config, which is really asking for all > fields. > >>>>> Are you sure you want that, that may also be slow. Try to ask for > just the > >>>>> fields you will be using. > >>>>> > >>>>> -- > >>>>> Jan Høydahl, search solution architect > >>>>> Cominvent AS - www.cominvent.com > >>>>> > >>>>>> 24. jan. 2019 kl. 14:59 skrev Zheng Lin Edwin Yeo < > >>>>> edwinye...@gmail.com>: > >>>>>> > >>>>>> Thanks for your reply. > >>>>>> > >>>>>> Below are what you have requested about our Solr setup, > configurations > >>>>>> files ,schema and results of debug queries: > >>>>>> > >>>>>> Looking forward to your advice and support on our problem. > >>>>>> > >>>>>> 1. System configurations > >>>>>> OS: Windows 10 Pro 64 bit > >>>>>> System Memory: 32GB > >>>>>> CPU: Intel(R) Core(TM) i7-4790K CPU @ 4.00GHz, 4 Core(s), 8 Logical > >>>>>> Processor(s) > >>>>>> HDD: 3.0 TB (free 2.1 TB) SATA > >>>>>> > >>>>>> 2. solrconfig.xml of customers and policies collection, and solr.in > >>>>> ,cmd > >>>>>> which can be download from the following link: > >>>>>> > >>>>> > https://drive.google.com/file/d/1AATjonQsEC5B0ldz27Xvx5A55Dp5ul8K/view?usp=sharing > >>>>>> > >>>>>> 3. The debug queries from both collections > >>>>>> > >>>>>> *3.1. Debug Query From Policies ( which is Slow)* > >>>>>> > >>>>>> "debug":{ > >>>>>> > >>>>>> "rawquerystring":"sherry", > >>>>>> > >>>>>> "querystring":"sherry", > >>>>>> > >>>>>> "parsedquery":"searchFields_tcs:sherry", > >>>>>> > >>>>>> "parsedquery_toString":"searchFields_tcs:sherry", > >>>>>> > >>>>>> "explain":{ > >>>>>> > >>>>>> "31702988":"\n14.540428 = weight(searchFields_tcs:sherry in > >>>>>> 3097315) [SchemaSimilarity], result of:\n 14.540428 = > >>>>>> score(doc=3097315,freq=5.0 = termFreq=5.0\n), product of:\n > >>>>>> 8.907154 = idf, computed as log(1 + (docCount - docFreq + 0.5) / > >>>>>> (docFreq + 0.5)) from:\n 812.0 = docFreq\n 6000000.0 = > >>>>>> docCount\n 1.6324438 = tfNorm, computed as (freq * (k1 + 1)) / > >>>>>> (freq + k1 * (1 - b + b * fieldLength / avgFieldLength)) from:\n > >>>>>> 5.0 = termFreq=5.0\n 1.2 = parameter k1\n 0.75 = parameter > >>>>>> b\n 19.397041 = avgFieldLength\n 31.0 = fieldLength\n”,.. > >>>>>> > >>>>>> "QParser":"LuceneQParser", > >>>>>> > >>>>>> "timing":{ > >>>>>> > >>>>>> "time":681.0, > >>>>>> > >>>>>> "prepare":{ > >>>>>> > >>>>>> "time":0.0, > >>>>>> > >>>>>> "query":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "facet":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "facet_module":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "mlt":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "highlight":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "stats":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "expand":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "terms":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "debug":{ > >>>>>> > >>>>>> "time":0.0}}, > >>>>>> > >>>>>> "process":{ > >>>>>> > >>>>>> "time":680.0, > >>>>>> > >>>>>> "query":{ > >>>>>> > >>>>>> "time":19.0}, > >>>>>> > >>>>>> "facet":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "facet_module":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "mlt":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "highlight":{ > >>>>>> > >>>>>> "time":651.0}, > >>>>>> > >>>>>> "stats":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "expand":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "terms":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "debug":{ > >>>>>> > >>>>>> "time":8.0}}, > >>>>>> > >>>>>> "loadFieldValues":{ > >>>>>> > >>>>>> "time":12.0}}}} > >>>>>> > >>>>>> > >>>>>> > >>>>>> *3.2. Debug Query From Customers (which is fast because we index it > >>>>> after > >>>>>> indexing Policies):* > >>>>>> > >>>>>> > >>>>>> > >>>>>> "debug":{ > >>>>>> > >>>>>> "rawquerystring":"sherry", > >>>>>> > >>>>>> "querystring":"sherry", > >>>>>> > >>>>>> "parsedquery":"searchFields_tcs:sherry", > >>>>>> > >>>>>> "parsedquery_toString":"searchFields_tcs:sherry", > >>>>>> > >>>>>> "explain":{ > >>>>>> > >>>>>> "S7900271B":"\n13.191501 = weight(searchFields_tcs:sherry in > >>>>>> 2453665) [SchemaSimilarity], result of:\n 13.191501 = > >>>>>> score(doc=2453665,freq=3.0 = termFreq=3.0\n), product of:\n > 9.08604 > >>>>>> = idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq + > >>>>>> 0.5)) from:\n 428.0 = docFreq\n 3784142.0 = docCount\n > >>>>>> 1.4518428 = tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 > - > >>>>>> b + b * fieldLength / avgFieldLength)) from:\n 3.0 = > >>>>>> termFreq=3.0\n 1.2 = parameter k1\n 0.75 = parameter b\n > >>>>>> 20.22558 = avgFieldLength\n 28.0 = fieldLength\n”, .. > >>>>>> > >>>>>> "QParser":"LuceneQParser", > >>>>>> > >>>>>> "timing":{ > >>>>>> > >>>>>> "time":38.0, > >>>>>> > >>>>>> "prepare":{ > >>>>>> > >>>>>> "time":1.0, > >>>>>> > >>>>>> "query":{ > >>>>>> > >>>>>> "time":1.0}, > >>>>>> > >>>>>> "facet":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "facet_module":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "mlt":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "highlight":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "stats":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "expand":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "terms":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "debug":{ > >>>>>> > >>>>>> "time":0.0}}, > >>>>>> > >>>>>> "process":{ > >>>>>> > >>>>>> "time":36.0, > >>>>>> > >>>>>> "query":{ > >>>>>> > >>>>>> "time":1.0}, > >>>>>> > >>>>>> "facet":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "facet_module":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "mlt":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "highlight":{ > >>>>>> > >>>>>> "time":31.0}, > >>>>>> > >>>>>> "stats":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "expand":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "terms":{ > >>>>>> > >>>>>> "time":0.0}, > >>>>>> > >>>>>> "debug":{ > >>>>>> > >>>>>> "time":3.0}}, > >>>>>> > >>>>>> "loadFieldValues":{ > >>>>>> > >>>>>> "time":13.0}}}} > >>>>>> > >>>>>> > >>>>>> > >>>>>> Best Regards, > >>>>>> Edwin > >>>>>> > >>>>>> On Thu, 24 Jan 2019 at 20:57, Jan Høydahl <jan....@cominvent.com> > >>>>> wrote: > >>>>>> > >>>>>>> It would be useful if you can disclose the machine configuration, > OS, > >>>>>>> memory, settings etc, as well as solr config including solr.in < > >>>>>>> http://solr.in/>.sh, solrconfig.xml etc, so we can see the whole > >>>>> picture > >>>>>>> of memory, GC, etc. > >>>>>>> You could also specify debugQuery=true on a slow search and check > the > >>>>>>> timings section for clues. What QTime are you seeing on the slow > >>>>> queries in > >>>>>>> solr.log? > >>>>>>> If that does not reveal the reason, I'd connect to your solr > >>>>> instance with > >>>>>>> a tool like jVisualVM or similar, to inspect what takes time. Or > >>>>> better, > >>>>>>> hook up to DataDog, SPM or some other cloud tool to get a full view > >>>>> of the > >>>>>>> system. > >>>>>>> > >>>>>>> -- > >>>>>>> Jan Høydahl, search solution architect > >>>>>>> Cominvent AS - www.cominvent.com > >>>>>>> > >>>>>>>> 24. jan. 2019 kl. 13:42 skrev Zheng Lin Edwin Yeo < > >>>>> edwinye...@gmail.com > >>>>>>>> : > >>>>>>>> > >>>>>>>> Hi Shawn, > >>>>>>>> > >>>>>>>> Unfortunately your reply of memory may not be valid. Please refer > >>>>> to my > >>>>>>>> explanation below of the strange behaviors (is it much more like a > >>>>> BUG > >>>>>>> than > >>>>>>>> anything else that is explainable): > >>>>>>>> > >>>>>>>> Note that we still have 18GB of free unused memory on the server. > >>>>>>>> > >>>>>>>> 1. We indexed the first collection called customers (3.7 millioin > >>>>> records > >>>>>>>> from CSV data), index size is 2.09GB. The search in customers for > >>>>> any > >>>>>>>> keyword is returned within 50ms (QTime) for using highlight > (unified > >>>>>>>> highlighter, posting, light term vectors) > >>>>>>>> > >>>>>>>> 2. Then we indexed the second collection called policies (6 > million > >>>>>>> records > >>>>>>>> from CSV data), index size is 2.55GB. The search in policies for > any > >>>>>>>> keyword is returned within 50ms (QTime) for using highlight > (unified > >>>>>>>> highlighter, posting, light term vectors) > >>>>>>>> > >>>>>>>> 3. But now any search in customers for any keywords (not from > cache) > >>>>>>> takes > >>>>>>>> as high as 1200ms (QTime). But still policies search remains very > >>>>> fast > >>>>>>>> (50ms). > >>>>>>>> > >>>>>>>> 4. So we decided to run the force optimize command on customers > >>>>>>> collection ( > >>>>>>>> > >>>>>>> > >>>>> > https://localhost:8983/edm/customers/update?optimize=true&numSegments=1&waitFlush=false > >>>>>>> ), > >>>>>>>> surprisingly after optimization the search on customers collection > >>>>> for > >>>>>>> any > >>>>>>>> keywords become very fast again (less than 50ms). BUT strangely, > the > >>>>>>> search > >>>>>>>> in policies collection become very slow (around 1200ms) without > any > >>>>>>> changes > >>>>>>>> to the policies collection. > >>>>>>>> > >>>>>>>> 5. Based on above result, we decided to run the force optimize > >>>>> command on > >>>>>>>> policies collection ( > >>>>>>>> > >>>>>>> > >>>>> > https://localhost:8983/edm/policies/update?optimize=true&numSegments=1&waitFlush=false > >>>>>>> ). > >>>>>>>> More surprisingly, after optimization the search on policies > >>>>> collection > >>>>>>> for > >>>>>>>> any keywords become very fast again (less than 50ms). BUT more > >>>>> strangely, > >>>>>>>> the search in customers collection again become very slow (around > >>>>> 1200ms) > >>>>>>>> without any changes to the customers collection. > >>>>>>>> > >>>>>>>> What a strange and unexpected behavior! If this is not a bug, how > >>>>> could > >>>>>>> you > >>>>>>>> explain the above very strange behavior in Solr 7.5. Could it be a > >>>>> bug? > >>>>>>>> > >>>>>>>> We would appreciate any support or help on our above situation. > >>>>>>>> > >>>>>>>> Thank you. > >>>>>>>> > >>>>>>>> Regards, > >>>>>>>> Edwin > >>>>>>>> > >>>>>>>> On Thu, 24 Jan 2019 at 16:14, Zheng Lin Edwin Yeo < > >>>>> edwinye...@gmail.com> > >>>>>>>> wrote: > >>>>>>>> > >>>>>>>>> Hi Shawn, > >>>>>>>>> > >>>>>>>>>> If the two collections have data on the same server(s), I can > see > >>>>> this > >>>>>>>>>> happening. More memory is consumed when there is additional > >>>>> data, and > >>>>>>>>>> when Solr needs more memory, performance might be affected. The > >>>>>>>>>> solution is generally to install more memory in the server. > >>>>>>>>> > >>>>>>>>> I have found that even after we delete the index in collection2, > >>>>> the > >>>>>>> query > >>>>>>>>> QTime for collection1 still remains slow. It does not goes back > to > >>>>> its > >>>>>>>>> previous fast speed before we index collection2. > >>>>>>>>> > >>>>>>>>> Regards, > >>>>>>>>> Edwin > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> On Thu, 24 Jan 2019 at 11:13, Zheng Lin Edwin Yeo < > >>>>> edwinye...@gmail.com > >>>>>>>> > >>>>>>>>> wrote: > >>>>>>>>> > >>>>>>>>>> Hi Shawn, > >>>>>>>>>> > >>>>>>>>>> Thanks for your reply. > >>>>>>>>>> > >>>>>>>>>> The log only shows a list the following and I don't see any > >>>>> other logs > >>>>>>>>>> besides these. > >>>>>>>>>> > >>>>>>>>>> 2019-01-24 02:47:57.925 INFO (qtp2131952342-1330) > [c:collectioin1 > >>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2] > >>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory > >>>>>>> update-script#processAdd: > >>>>>>>>>> id=13245417 > >>>>>>>>>> 2019-01-24 02:47:57.957 INFO (qtp2131952342-1330) > [c:collectioin1 > >>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2] > >>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory > >>>>>>> update-script#processAdd: > >>>>>>>>>> id=13245430 > >>>>>>>>>> 2019-01-24 02:47:57.957 INFO (qtp2131952342-1330) > [c:collectioin1 > >>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2] > >>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory > >>>>>>> update-script#processAdd: > >>>>>>>>>> id=13245435 > >>>>>>>>>> > >>>>>>>>>> There is no change to the segments info. but the slowdown in the > >>>>> first > >>>>>>>>>> collection is very drastic. > >>>>>>>>>> Before the indexing of collection2, the collection1 query QTime > >>>>> are in > >>>>>>>>>> the range of 4 to 50 ms. However, after indexing collection2, > the > >>>>>>>>>> collection1 query QTime increases to more than 1000 ms. The > index > >>>>> are > >>>>>>> done > >>>>>>>>>> in CSV format, and the size of the index is 3GB. > >>>>>>>>>> > >>>>>>>>>> Regards, > >>>>>>>>>> Edwin > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> On Thu, 24 Jan 2019 at 01:09, Shawn Heisey <apa...@elyograg.org > > > >>>>>>> wrote: > >>>>>>>>>> > >>>>>>>>>>> On 1/23/2019 10:01 AM, Zheng Lin Edwin Yeo wrote: > >>>>>>>>>>>> I am using Solr 7.5.0, and currently I am facing an issue of > >>>>> when I > >>>>>>> am > >>>>>>>>>>>> indexing in collection2, the indexing affects the records in > >>>>>>>>>>> collection1. > >>>>>>>>>>>> Although the records are still intact, it seems that the > >>>>> settings of > >>>>>>>>>>> the > >>>>>>>>>>>> termVecotrs get wipe out, and the index size of collection1 > >>>>> reduced > >>>>>>>>>>> from > >>>>>>>>>>>> 3.3GB to 2.1GB after I do the indexing in collection2. > >>>>>>>>>>> > >>>>>>>>>>> This should not be possible. Indexing in one collection should > >>>>> have > >>>>>>>>>>> absolutely no effect on another collection. > >>>>>>>>>>> > >>>>>>>>>>> If logging has been left at its default settings, the solr.log > >>>>> file > >>>>>>>>>>> should have enough info to show what actually happened. > >>>>>>>>>>> > >>>>>>>>>>>> Also, the search in > >>>>>>>>>>>> collection1, which was originall very fast, becomes very slow > >>>>> after > >>>>>>> the > >>>>>>>>>>>> indexing is done is collection2. > >>>>>>>>>>> > >>>>>>>>>>> If the two collections have data on the same server(s), I can > >>>>> see this > >>>>>>>>>>> happening. More memory is consumed when there is additional > >>>>> data, and > >>>>>>>>>>> when Solr needs more memory, performance might be affected. > The > >>>>>>>>>>> solution is generally to install more memory in the server. If > >>>>> the > >>>>>>>>>>> system is working, there should be no need to increase the heap > >>>>> size > >>>>>>>>>>> when the memory size increases ... but there can be situations > >>>>> where > >>>>>>> the > >>>>>>>>>>> heap is a little bit too small, where you WOULD want to > increase > >>>>> the > >>>>>>>>>>> heap size. > >>>>>>>>>>> > >>>>>>>>>>> Thanks, > >>>>>>>>>>> Shawn > >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>> > >>>>>>> > >>>>> > >>>>> > >