On Tue, 2007-09-25 at 10:03 +0530, Roopesh P Raj wrote:

DISCLAIMER:
Please, I am subscribed to the user list and there is no need to write
me directly nor cc me in your response. More since we are an open source
project off-list communication is suboptimal and harmful to the
community. The community has many eyes which can see possible problems
with some solution and propose better ones. Further the mailing list has
an archive and proofed solution can be searched. If we all share
off-list mailings no solutions go into the archive and we always have to
repeat the same mails.

PLEASE write to the ml!

> > http://wiki.apache.org/solr/CommonQueryParameters
> 
> > It should work as described in the above document (with the start
> > parameter.
> 
> > e.g. 
> > data = c.search(q='query', fl='id score unique_id Message-ID To From
> > Subject',rows=50, wt='python',start=50)
> 
> > HTH
> > --
> 
> Hi,
> 
> I my application there is a provision to copy the archive based on date 
> indexed. 
> In this case the number of search results may exceed the high number I have 
> assigned to rows, say rows=10000000. I wanted to avoid this situation. In 
> this 
> situation I don't want paginated queries. 
> 
> Can you please tell me how to approach this particular situation.

I think the best way is to
1) get the first response document  (rows=50,start=0)
2) parse the response to see how many results you have
3) do a loop (rows=50,start=50*x) and call solr till you have all
results.

Like Jérôme stated:
On Mon, 2007-09-24 at 12:45 +0100, Jérôme Etévé wrote:
> By design, it's not very efficient to ask for a large number of
> results with solr/lucene. I think you will face performance and memory
> problems if you do that. 

HTH

salu2
-- 
Thorsten Scherler                                 thorsten.at.apache.org
Open Source Java                      consulting, training and solutions

Reply via email to