> My data is in an enormous text file that is parsed in python,
You mean it is in Python s-expressions? I don't think there is a
parser in DIH for that.
On Thu, Jul 19, 2012 at 9:27 AM, Erick Erickson wrote:
> First, turn off all your soft commit stuff, that won't help in your situation.
> If yo
First, turn off all your soft commit stuff, that won't help in your situation.
If you do leave autocommit on, make it a really high number
(let's say 1,000,000 to start).
You won't have to make 300M calls, you can batch, say, 1,000 docs
into each request.
DIH supports a bunch of different data so
Hi Jonatan,
Ideally you'd use a Solr API client that allowed batched updates, so
you'd be sending documents 100 at a time, say. Alternatively, if
you're good with Java, you could build an index by using the
EmbeddedSolrServer class in the same process as the code you use to
parse the documents. Bu