On Tue, 2007-09-25 at 10:03 +0530, Roopesh P Raj wrote: DISCLAIMER: Please, I am subscribed to the user list and there is no need to write me directly nor cc me in your response. More since we are an open source project off-list communication is suboptimal and harmful to the community. The community has many eyes which can see possible problems with some solution and propose better ones. Further the mailing list has an archive and proofed solution can be searched. If we all share off-list mailings no solutions go into the archive and we always have to repeat the same mails.
PLEASE write to the ml! > > http://wiki.apache.org/solr/CommonQueryParameters > > > It should work as described in the above document (with the start > > parameter. > > > e.g. > > data = c.search(q='query', fl='id score unique_id Message-ID To From > > Subject',rows=50, wt='python',start=50) > > > HTH > > -- > > Hi, > > I my application there is a provision to copy the archive based on date > indexed. > In this case the number of search results may exceed the high number I have > assigned to rows, say rows=10000000. I wanted to avoid this situation. In > this > situation I don't want paginated queries. > > Can you please tell me how to approach this particular situation. I think the best way is to 1) get the first response document (rows=50,start=0) 2) parse the response to see how many results you have 3) do a loop (rows=50,start=50*x) and call solr till you have all results. Like Jérôme stated: On Mon, 2007-09-24 at 12:45 +0100, Jérôme Etévé wrote: > By design, it's not very efficient to ask for a large number of > results with solr/lucene. I think you will face performance and memory > problems if you do that. HTH salu2 -- Thorsten Scherler thorsten.at.apache.org Open Source Java consulting, training and solutions