On Fri, Aug 01, 2008 at 03:36:13PM -0400, Ian Connor wrote:
> I have a number of documents in files
>
> 1.xml
> 2.xml
> ...
> 17M.xml
>
> I have been using cat to join them all together:
>
> cat 1.xml 2.xml ... 1000.xml | grep -v '<\/add>' > /tmp/post.xml
>
> and posting them with cur
ze of your documents and the
analysis being done on them.
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Message
> From: Ian Connor <[EMAIL PROTECTED]>
> To: solr-user@lucene.apache.org
> Sent: Friday, August 1, 2008 5:08:00 PM
> Subject: Re: f
en only 100 docs/second.
>
> Otis
> --
> Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
>
>
>
> - Original Message
>> From: Ian Connor <[EMAIL PROTECTED]>
>> To: solr-user@lucene.apache.org
>> Sent: Friday, August 1, 2008 3:36:1
ssage
> From: Ian Connor <[EMAIL PROTECTED]>
> To: solr-user@lucene.apache.org
> Sent: Friday, August 1, 2008 3:36:13 PM
> Subject: fastest way to load documents
>
> I have a number of documents in files
>
> 1.xml
> 2.xml
> ...
> 17M.xml
>
&g
I have a number of documents in files
1.xml
2.xml
...
17M.xml
I have been using cat to join them all together:
cat 1.xml 2.xml ... 1000.xml | grep -v '<\/add>' > /tmp/post.xml
and posting them with curl:
curl -d @/tmp/post.xml 'http://localhost:8983/solr/update' -H
'Content-Type: tex