I think your biggest problem is requesting 70,000 records from Solr.
That is not going to be fast.

Two suggestions:

1. Use paging. Get the results in chunks, 10, 25, 100, whatever.

2. Since you are running out of memory parsing XML, I'm guessing
that you're using a DOM-style parser. Don't do that. You do not
need to create elaborate structures, strip mine the data, then
throw those structures away. Instead, us a streaming parser, like Stax.

The sounds like an XY problem. What are you trying to achieve by
fetching 10,000 records? There is probably a better way to do it.

wunder


On 12/12/07 11:58 AM, "Owens, Martin" <[EMAIL PROTECTED]> wrote:

> Hello everyone,
> 
> I'm looking for a better solution that the current xml output we're currently
> getting; if you return more than 70k records the webserver can no longer cope
> with parsing the xml and the machine falls over out of memory.
> 
> Ideally what we'd like is for the search results to go directly into a
> temporary mysql table so we can link against it in a further request from the
> web server. Does anyone know any plugs or people who have done anything along
> these lines?
> 
> We might be able to settle for receiving the single field column as a csv type
> file, that would at least let us cut down on the processing and parsing. I see
> there is a csv indexer but do we have a csv output plugin?
> 
> Once again thank you all for your help.
> 
> Best Regards, Martin Ownes

Reply via email to