On 14-Sep-07, at 5:19 AM, Thompson,Roger wrote:
Hi there!
I am embarking on re-engineering an application using Solr/Lucene (If
you'd like to see the current manifestation go to:
fictionfinder.oclc.org). The database for this application
consists of
approximatly 1.4 million records of varying size for the "work"
record,
and another database of 1.9 million bibliographic records. I fear
that
loading this through http will take several days, perhaps a week. Do
any of you have a way to do a large batch load of the DB?
I can index 2 million web documents in 7 hours over http. Just batch
a few (10) docs per http POST, and use around N+1 threads (N=#
processors).
-Mike