Actually I still have this error : " No read Solr server available "
sunnyfr wrote:
>
> Ok sorry I just add the parameter -XX:+UseParallelGC and it seems to don't
> go oom.
>
>
>
>
> sunnyfr wrote:
>>
>> Actually I just notices, lot of request didn"t bring back correct answer,
>> but " No read Solr server available" so my jmeter didn't take that for an
>> error. Obviously out of memory, and a file gc.log is created with :
>> 0.054: [GC [PSYoungGen: 5121K->256K(298688K)] 5121K->256K(981376K),
>> 0.0020630 secs] [Times: user=0.00 sys=0.00, real=0.00 secs]
>> 0.056: [Full GC (System) [PSYoungGen: 256K->0K(298688K)] [PSOldGen:
>> 0K->180K(682688K)] 256K->180K(981376K) [PSPermGen: 3002K->3002K(21248K)],
>> 0.0055170 secs] [Times: user=0.00 sys=0.00, real=0.00 secs]
>>
>> so far my tomcat55 file is configurate like that :
>> JAVA_OPTS="-Xms1000m -Xmx4000m -XX:+HeapDumpOnOutOfMemoryError
>> -Xloggc:gc.log -XX:+PrintGCDetails -XX:+PrintGCTimeStamps"
>>
>> My error:
>> Dec 11 14:16:27 solr-test jsvc.exec[30653]: Dec 11, 2008 2:16:27 PM
>> org.apache.solr.core.SolrCore execute INFO: [video] webapp=/solr
>> path=/admin/ping params={} hits=0 status=0 QTime=1 Dec 11, 2008 2:16:27
>> PM org.apache.solr.core.SolrCore execute INFO: [video] webapp=/solr
>> path=/admin/ping params={} status=0 QTime=1 Dec 11, 2008 2:16:27 PM
>> org.apache.solr.core.SolrCore execute INFO: [video] webapp=/solr
>> path=/admin/ping params={} hits=0 status=0 QTime=1 Dec 11, 2008 2:16:27
>> PM org.apache.solr.core.SolrCore execute INFO: [video] webapp=/solr
>> path=/admin/ping params={} status=0 QTime=2
>> Dec 11 14:16:27 solr-test jsvc.exec[30653]: java.lang.OutOfMemoryError:
>> GC overhead limit exceeded Dumping heap to java_pid30655.hprof ...
>>
>>
>>
>>
>> Thanks for your help
>>
>>
>>
>> Shalin Shekhar Mangar wrote:
>>>
>>> Are each of those queries unique?
>>>
>>> First time queries are slower. They are cached by Solr and the same
>>> query
>>> again will return results very quickly because it won't need to hit the
>>> file
>>> system.
>>>
>>> On Thu, Dec 11, 2008 at 4:08 PM, sunnyfr <[EMAIL PROTECTED]> wrote:
>>>
>>>>
>>>> Hi,
>>>>
>>>> I'm doing a stress test on solr.
>>>> I've around 8,5M of doc, the size of my data's directory is 5,6G.
>>>>
>>>> I've indexed again my data to make it faster, and applied all the last
>>>> patch.
>>>> My index data store just two field : id and text (which is a copy of
>>>> three
>>>> fiels)
>>>> But I still think it's very long, what do you think?
>>>>
>>>> For 50request/sec during 40mn, my average respond time : 1235msec.
>>>> 49430request.
>>>>
>>>> When I make this test with 100request second during 10mn and 10 other
>>>> minutes with 50 request : my average respond time is 1600msec. Don't
>>>> you
>>>> think it's a bit long.
>>>>
>>>> Should I partition this index more ? or what should I do to make this
>>>> work
>>>> faster.
>>>> I can read post with people who have just 300msec request for 300Go of
>>>> index
>>>> partionned ?
>>>> My request for collecting all this book is quite complex and have lot
>>>> of
>>>> table linked together, maybe it would be faster if I create a csv file
>>>> ?
>>>>
>>>> The server that I'm using for the test has 8G of memory.
>>>> 4CPU : Intel(R) Xeon(R) CPU 5160 @ 3.00GHz
>>>> Tomcat55 : -Xms2000m -Xmx4000m
>>>> Solr 1.3.
>>>>
>>>> What can I modify to make it more performant ? memory, indexation ...?
>>>> Does it can come from my request to the mysql database which is too
>>>> much
>>>> linked ?
>>>>
>>>> Thanks a lot for your help,
>>>> Johanna
>>>>
>>>>
>>>> --
>>>> View this message in context:
>>>> http://www.nabble.com/Make-it-more-performant---solr-1.3---1200msec-respond-time.-tp20953079p20953079.html
>>>> Sent from the Solr - User mailing list archive at Nabble.com.
>>>>
>>>>
>>>
>>>
>>> --
>>> Regards,
>>> Shalin Shekhar Mangar.
>>>
>>>
>>
>>
>
>
--
View this message in context:
http://www.nabble.com/Make-it-more-performant---solr-1.3---1200msec-respond-time.-tp20953079p20955609.html
Sent from the Solr - User mailing list archive at Nabble.com.