Hi Walter,

sort of shameless plug - I ran into similar issues and wrote a JMeter SLA 
Reporting Backend - https://github.com/sgoeschl/jmeter-sla-report 
<https://github.com/sgoeschl/jmeter-sla-report>

* It reads the CSV/XML JMeter report file and sorts the response times in 
logarithmic buckets 
* the XML processor uses a Stax parser to handle huge JTL files (exceeding 1 GB)
* it also caters for merging JTL files when running multiple JMeter instances

Cheers,

Siegfried Goeschl



> On 06 Apr 2015, at 22:57, Walter Underwood <wun...@wunderwood.org> wrote:
> 
> The load testing is the easiest part.
> 
> We use JMeter to replay the prod logs. We start about a hundred threads and 
> use ConstantThroughputTimer to control the traffic level. JMeter tends to 
> fall over with two much data graphing, so we run it headless. Then we post 
> process with JMeter Plugins to get percentiles.
> 
> The complicated part of the servlet filter was getting it configured in 
> Tomcat. The code itself is not too bad.
> 
> wunder
> Walter Underwood
> wun...@wunderwood.org
> http://observer.wunderwood.org/  (my blog)
> 
> On Apr 6, 2015, at 1:49 PM, Siegfried Goeschl <sgoes...@gmx.at> wrote:
> 
>> The good-sounding thing - you can do that easily with JMeter running the GUI 
>> or the command-line
>> 
>> Cheers,
>> 
>> Siegfried Goeschl
>> 
>>> On 06 Apr 2015, at 21:35, Davis, Daniel (NIH/NLM) [C] 
>>> <daniel.da...@nih.gov> wrote:
>>> 
>>> This sounds really good:
>>> 
>>> "For load testing, we replay production logs to test that we meet the SLA 
>>> at a given traffic level."
>>> 
>>> The rest sounds complicated.   Ah well, that's the job.
>>> 
>>> -----Original Message-----
>>> From: Walter Underwood [mailto:wun...@wunderwood.org] 
>>> Sent: Monday, April 06, 2015 2:48 PM
>>> To: solr-user@lucene.apache.org
>>> Subject: Re: Measuring QPS
>>> 
>>> We built a servlet request filter that is configured in front of the Solr 
>>> servlets. It reports response times to metricsd, using the Codahale library.
>>> 
>>> That gives us counts, rates, and response time metrics. We mostly look at 
>>> percentiles, because averages are thrown off by outliers. Average is just 
>>> the wrong metric for a one-sided distribution like response times.
>>> 
>>> We use Graphite to display the 95th percentile response time for each 
>>> request handler. We use Tattle for alerting on those metrics.
>>> 
>>> We also use New Relic for a different look at the performance. It is good 
>>> at tracking from the front end through to Solr.
>>> 
>>> For load testing, we replay production logs to test that we meet the SLA at 
>>> a given traffic level.
>>> 
>>> Walter Underwood
>>> wun...@wunderwood.org
>>> http://observer.wunderwood.org/  (my blog)
>>> 
>>> On Apr 6, 2015, at 11:31 AM, Davis, Daniel (NIH/NLM) [C] 
>>> <daniel.da...@nih.gov> wrote:
>>> 
>>>> OK,
>>>> 
>>>> I have a lot of chutzpah posting that here ;)    The other guys answering 
>>>> the questions can probably explain it better.
>>>> I love showing off, however, so please forgive me.
>>>> 
>>>> -----Original Message-----
>>>> From: Davis, Daniel (NIH/NLM) [C]
>>>> Sent: Monday, April 06, 2015 2:25 PM
>>>> To: solr-user@lucene.apache.org
>>>> Subject: RE: Measuring QPS
>>>> 
>>>> Its very common to do autocomplete based on popular queries/titles over 
>>>> some sliding time window.   Some enterprise search systems even apply age 
>>>> weighting so that they don't need to re-index but continuously add to the 
>>>> index.   This way, they can do autocomplete based on what's popular these 
>>>> days.
>>>> 
>>>> We use relevance/field boosts/phrase matching etc. to get the best guess 
>>>> about what results they want to see.   This is similar - we use relevance, 
>>>> field boosting to guess what users want to search for.   Zipf's law 
>>>> applies to searches as well as results.
>>>> 
>>>> -----Original Message-----
>>>> From: Siegfried Goeschl [mailto:sgoes...@gmx.at]
>>>> Sent: Monday, April 06, 2015 2:17 PM
>>>> To: solr-user@lucene.apache.org
>>>> Subject: Re: Measuring QPS
>>>> 
>>>> Hi Daniel,
>>>> 
>>>> interesting - I never thought of autocompletion but for keeping track 
>>>> of user behaviour :-)
>>>> 
>>>> * the numbers are helpful for the online advertisement team to sell 
>>>> campaigns
>>>> * it is used for sanity checks - sensible queries returning no results 
>>>> or returning too many results
>>>> 
>>>> Cheers,
>>>> 
>>>> Siegfried Goeschl
>>>> 
>>>>> On 06 Apr 2015, at 20:04, Davis, Daniel (NIH/NLM) [C] 
>>>>> <daniel.da...@nih.gov> wrote:
>>>>> 
>>>>> Siegfried,
>>>>> 
>>>>> It is early days as yet.   I don't think we need a code drop.   AFAIK, 
>>>>> none of our current Solr applications autocomplete the search box based 
>>>>> on popular query/title keywords.   We have other applications that do 
>>>>> that, but they don't use Solr.
>>>>> 
>>>>> Thanks again,
>>>>> 
>>>>> Dan
>>>>> 
>>>>> -----Original Message-----
>>>>> From: Siegfried Goeschl [mailto:sgoes...@gmx.at]
>>>>> Sent: Monday, April 06, 2015 1:42 PM
>>>>> To: solr-user@lucene.apache.org
>>>>> Subject: Re: Measuring QPS
>>>>> 
>>>>> Hi Dan,
>>>>> 
>>>>> at willhaben.at (customer of mine) two SOLR components were written 
>>>>> for SOLR 3 and ported to SORL 4
>>>>> 
>>>>> 1) SlowQueryLog which dumps long-running search requests into a log 
>>>>> file
>>>>> 
>>>>> 2) Most Frequent Search Terms allowing to query & filter the most 
>>>>> frequent user search terms over the browser
>>>>> 
>>>>> Some notes along the line
>>>>> 
>>>>> 
>>>>> * For both components I have the "GO" to open source them but I never 
>>>>> had enough time to do that (shame on me) - see
>>>>> https://issues.apache.org/jira/browse/SOLR-4056
>>>>> 
>>>>> * The Most Frequent Search Term component actually mimics a SOLR 
>>>>> server you feed the user search terms so this might be a better 
>>>>> solution in the long run. But this requires to have a separate SOLR 
>>>>> core & ingest  plus GUI (check out SILK or ELK) - in other words more 
>>>>> moving parts in production :-)
>>>>> 
>>>>> * If there is sufficient interest I can make a code drop on GitHub
>>>>> 
>>>>> Cheers,
>>>>> 
>>>>> Siegfried Goeschl
>>>>> 
>>>>> 
>>>>> 
>>>>>> On 06 Apr 2015, at 16:25, Davis, Daniel (NIH/NLM) [C] 
>>>>>> <daniel.da...@nih.gov> wrote:
>>>>>> 
>>>>>> Siegfried,
>>>>>> 
>>>>>> This is a wonderful find.   The second presentation is a nice write-up 
>>>>>> of a large number of free tools.   The first presentation prompts a 
>>>>>> question - did you add custom request handlers/code to automate 
>>>>>> determination of best user search terms?   Did any of your custom work 
>>>>>> end-up in Solr?
>>>>>> 
>>>>>> Thank you so much,
>>>>>> 
>>>>>> Dan
>>>>>> 
>>>>>> P.S. - your first presentation takes me back to seeing "Angrif der 
>>>>>> Klonkrieger" in Berlin after a conference - Hayden Christensen was less 
>>>>>> annoying in German, because my wife and I don't speak German ;)   I 
>>>>>> haven't thought of that in a while.
>>>>>> 
>>>>>> -----Original Message-----
>>>>>> From: Siegfried Goeschl [mailto:sgoes...@gmx.at]
>>>>>> Sent: Saturday, April 04, 2015 4:54 AM
>>>>>> To: solr-user@lucene.apache.org
>>>>>> Subject: Re: Measuring QPS
>>>>>> 
>>>>>> Hi Dan,
>>>>>> 
>>>>>> I'm using JavaMelody for my SOLR production servers - gives you the 
>>>>>> relevant HTTP stats (what's happening now & historical data) plus 
>>>>>> JVM monitoring as additional benefit. The servers are deployed on 
>>>>>> Tomcat so I'm of little help regarding Jetty - having said that
>>>>>> 
>>>>>> * you need two Jars (javamelody & robin)
>>>>>> * tinker with web.xml
>>>>>> 
>>>>>> Here are two of my presentations mentioning JavaMelody (plus some 
>>>>>> other stuff)
>>>>>> 
>>>>>> http://people.apache.org/~sgoeschl/presentations/solr-from-developme
>>>>>> n
>>>>>> t
>>>>>> -to-production-20121210.pdf
>>>>>> <http://people.apache.org/~sgoeschl/presentations/solr-from-developm
>>>>>> e
>>>>>> n
>>>>>> t-to-production-20121210.pdf>
>>>>>> http://people.apache.org/~sgoeschl/presentations/jsug-2015/jee-perfo
>>>>>> r
>>>>>> m
>>>>>> ance-monitoring.pdf
>>>>>> <http://people.apache.org/~sgoeschl/presentations/jsug-2015/jee-perf
>>>>>> o
>>>>>> r
>>>>>> mance-monitoring.pdf>
>>>>>> 
>>>>>> Cheers,
>>>>>> 
>>>>>> Siegfried Goeschl
>>>>>> 
>>>>>>> On 03 Apr 2015, at 17:53, Shawn Heisey <apa...@elyograg.org> wrote:
>>>>>>> 
>>>>>>> On 4/3/2015 9:37 AM, Davis, Daniel (NIH/NLM) [C] wrote:
>>>>>>>> I wanted to gather QPS for our production Solr instances, but I was 
>>>>>>>> surprised that the Admin UI did not contain this information.   We are 
>>>>>>>> running a mix of versions, but mostly 4.10 at this point.   We are not 
>>>>>>>> using SolrCloud at present; that's part of why I'm checking - I want 
>>>>>>>> to validate the size of our existing setup and what sort of SolrCloud 
>>>>>>>> setup would be needed to centralize several of them.
>>>>>>>> 
>>>>>>>> What is the best way to gather QPS information?
>>>>>>>> 
>>>>>>>> What is the best way to add information like this to the Admin UI, if 
>>>>>>>> I decide to take that step?
>>>>>>> 
>>>>>>> As of Solr 4.1 (three years ago), request rate information is 
>>>>>>> available in the admin UI and via JMX.  In the admin UI, choose a 
>>>>>>> core from the dropdown, click on Plugins/Stats, then QUERYHANDLER, 
>>>>>>> and open the handler you wish to examine.  You have 
>>>>>>> avgRequestsPerSecond, which is calculated for the entire runtime of 
>>>>>>> the SolrCore, as well as 5minRateReqsPerSecond and 
>>>>>>> 15minRateReqsPerSecond, which are far more useful pieces of information.
>>>>>>> 
>>>>>>> https://issues.apache.org/jira/browse/SOLR-1972
>>>>>>> 
>>>>>>> Thanks,
>>>>>>> Shawn
>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> 
> 

Reply via email to