If I don't do highlighting, it's really fast. Optimize has no effect. -Peter
On Jul 29, 2010, at 11:54 AM, dc tech wrote: > Are you storing the entire log file text in SOLR? That's almost 3gb of > text that you are storing in the SOLR. Try to > 1) Is this first time performance or on repaat queries with the same fields? > 2) Optimze the index and test performance again > 3) index without storing the text and see what the performance looks like. > > > On 7/29/10, Peter Spam <ps...@mac.com> wrote: >> Any ideas? I've got 5000 documents with an average size of 850k each, and >> it sometimes takes 2 minutes for a query to come back when highlighting is >> turned on! Help! >> >> >> -Pete >> >> On Jul 21, 2010, at 2:41 PM, Peter Spam wrote: >> >>> From the mailing list archive, Koji wrote: >>> >>>> 1. Provide another field for highlighting and use copyField to copy >>>> plainText to the highlighting field. >>> >>> and Lance wrote: >>> http://www.mail-archive.com/solr-user@lucene.apache.org/msg35548.html >>> >>>> If you want to highlight field X, doing the >>>> termOffsets/termPositions/termVectors will make highlighting that field >>>> faster. You should make a separate field and apply these options to that >>>> field. >>>> >>>> Now: doing a copyfield adds a "value" to a multiValued field. For a text >>>> field, you get a multi-valued text field. You should only copy one value >>>> to the highlighted field, so just copyField the document to your special >>>> field. To enforce this, I would add multiValued="false" to that field, >>>> just to avoid mistakes. >>>> >>>> So, all_text should be indexed without the term* attributes, and should >>>> not be stored. Then your document stored in a separate field that you use >>>> for highlighting and has the term* attributes. >>> >>> I've been experimenting with this, and here's what I've tried: >>> >>> <field name="body" type="text_pl" indexed="true" stored="false" >>> multiValued="true" termVectors="true" termPositions="true" termOff >>> sets="true" /> >>> <field name="body_all" type="text_pl" indexed="false" stored="true" >>> multiValued="true" /> >>> <copyField source="body" dest="body_all"/> >>> >>> ... but it's still very slow (10+ seconds). Why is it better to have two >>> fields (one indexed but not stored, and the other not indexed but stored) >>> rather than just one field that's both indexed and stored? >>> >>> >>> From the Perf wiki page http://wiki.apache.org/solr/SolrPerformanceFactors >>> >>>> If you aren't always using all the stored fields, then enabling lazy >>>> field loading can be a huge boon, especially if compressed fields are >>>> used. >>> >>> What does this mean? How do you load a field lazily? >>> >>> Thanks for your time, guys - this has started to become frustrating, since >>> it works so well, but is very slow! >>> >>> >>> -Pete >>> >>> On Jul 20, 2010, at 5:36 PM, Peter Spam wrote: >>> >>>> Data set: About 4,000 log files (will eventually grow to millions). >>>> Average log file is 850k. Largest log file (so far) is about 70MB. >>>> >>>> Problem: When I search for common terms, the query time goes from under >>>> 2-3 seconds to about 60 seconds. TermVectors etc are enabled. When I >>>> disable highlighting, performance improves a lot, but is still slow for >>>> some queries (7 seconds). Thanks in advance for any ideas! >>>> >>>> >>>> -Peter >>>> >>>> >>>> ------------------------------------------------------------------------------------------------------------------------------------- >>>> >>>> 4GB RAM server >>>> % java -Xms2048M -Xmx3072M -jar start.jar >>>> >>>> ------------------------------------------------------------------------------------------------------------------------------------- >>>> >>>> schema.xml changes: >>>> >>>> <fieldType name="text_pl" class="solr.TextField"> >>>> <analyzer> >>>> <tokenizer class="solr.WhitespaceTokenizerFactory"/> >>>> <filter class="solr.LowerCaseFilterFactory"/> >>>> <filter class="solr.WordDelimiterFilterFactory" generateWordParts="0" >>>> generateNumberParts="0" catenateWords="0" catenateNumbers="0" >>>> catenateAll="0" splitOnCaseChange="0"/> >>>> </analyzer> >>>> </fieldType> >>>> >>>> ... >>>> >>>> <field name="body" type="text_pl" indexed="true" stored="true" >>>> multiValued="false" termVectors="true" termPositions="true" >>>> termOffsets="true" /> >>>> <field name="timestamp" type="date" indexed="true" stored="true" >>>> default="NOW" multiValued="false"/> >>>> <field name="version" type="string" indexed="true" stored="true" >>>> multiValued="false"/> >>>> <field name="device" type="string" indexed="true" stored="true" >>>> multiValued="false"/> >>>> <field name="filename" type="string" indexed="true" stored="true" >>>> multiValued="false"/> >>>> <field name="filesize" type="long" indexed="true" stored="true" >>>> multiValued="false"/> >>>> <field name="pversion" type="int" indexed="true" stored="true" >>>> multiValued="false"/> >>>> <field name="first2md5" type="string" indexed="false" stored="true" >>>> multiValued="false"/> >>>> <field name="ckey" type="string" indexed="true" stored="true" >>>> multiValued="false"/> >>>> >>>> ... >>>> >>>> <dynamicField name="*" type="ignored" multiValued="true" /> >>>> <defaultSearchField>body</defaultSearchField> >>>> <solrQueryParser defaultOperator="AND"/> >>>> >>>> ------------------------------------------------------------------------------------------------------------------------------------- >>>> >>>> solrconfig.xml changes: >>>> >>>> <maxFieldLength>2147483647</maxFieldLength> >>>> <ramBufferSizeMB>128</ramBufferSizeMB> >>>> >>>> ------------------------------------------------------------------------------------------------------------------------------------- >>>> >>>> The query: >>>> >>>> rowStr = "&rows=10" >>>> facet = >>>> "&facet=true&facet.limit=10&facet.field=device&facet.field=ckey&facet.field=version" >>>> fields = "&fl=id,score,filename,version,device,first2md5,filesize,ckey" >>>> termvectors = "&tv=true&qt=tvrh&tv.all=true" >>>> hl = "&hl=true&hl.fl=body&hl.snippets=1&hl.fragsize=400" >>>> regexv = "(?m)^.*\n.*\n.*$" >>>> hl_regex = "&hl.regex.pattern=" + CGI::escape(regexv) + >>>> "&hl.regex.slop=1&hl.fragmenter=regex&hl.regex.maxAnalyzedChars=2147483647&hl.maxAnalyzedChars=2147483647" >>>> justq = '&q=' + CGI::escape('body:' + fuzzy + p['q'].to_s.gsub(/\\/, >>>> '').gsub(/([:~!<>="])/,'\\\\\1') + fuzzy + minLogSizeStr) >>>> >>>> thequery = '/solr/select?timeAllowed=5000&wt=ruby' + (p['fq'].empty? ? '' >>>> : ('&fq='+p['fq'].to_s) ) + justq + rowStr + facet + fields + termvectors >>>> + hl + hl_regex >>>> >>>> baseurl = '/cgi-bin/search.rb?q=' + CGI::escape(p['q'].to_s) + '&rows=' + >>>> p['rows'].to_s + '&minLogSize=' + p['minLogSize'].to_s >>>> >>> >> >> > > -- > Sent from my mobile device