Hey Shawn / Solr User Group,
 
This makes perfect sense to me. Thanks for the thorough answer.  
     "The CSV update handler works at a lower level than the DataImport 
handler, and doesn't 
have "clean" or "full-import" options, which defaults to clean=true. The DIH is 
like a full application embedded inside Solr, one that uses 
an update handler -- it is not itself an update handler.  When clean=true or 
using full-import without a clean option, DIH itself sends 
a "delete all documents" update request."
 
And similiarly, my assumption is in the event of a non-syntactical 
failure/interuption  (such as a server crash) during the CSV Update a rollback 
(stream.body=<rollback/>) would also need to be manually requested (or 
automatted but outside of Solr) where as the DIH automates this Request on my 
behalf as well...? Is there anyway to detect this failure or interuption?...A 
real example is, I was in the process of indexing data via the CSV Update and 
somebody bounced the server before it completed. No actual errors were produced 
but it appeared that the CSV Update process stopped at the point of the reboot. 
My assumption is, if I had passed in a rollback, I'd get the previously indexed 
data , given I didn't request a delete beforehand (haven't yet tested this). 
But wondering, how I could automatically detect this? This I guess is where DIH 
starts gaining some merit. Also - the response that the DIH produces when the 
indexing process is complete appears
 to be a lot more mature in that it explicity suggest the index completed and 
that information can can be re-queried. It would be nice if the CSV Update 
provided a similiar response..my assumption is it would first need to know how 
many lines exist on the file in order to know whether or not the job actually 
completed...
 
 Also - outside of solr initiating a delete due to encountering the same 
UniqueKey, is there anything else that could cause a delete to be initiated by 
Solr? 

Lastly, is there any concern of running multiple Update CSV requests on 
different data files containing different data? 

Thanks in advance. This was very helpful.

Mike
 

________________________________
From: Shawn Heisey <s...@elyograg.org>
To: solr-user@lucene.apache.org 
Sent: Monday, July 1, 2013 2:30 PM
Subject: Re: FileDataSource vs JdbcDataSouce (speed) Solr 3.5


On 7/1/2013 12:56 PM, Mike L. wrote:
>  Hey Ahmet / Solr User Group,
>
>    I tried using the built in UpdateCSV and it runs A LOT faster than a 
>FileDataSource DIH as illustrated below. However, I am a bit confused about 
>the numDocs/maxDoc values when doing an import this way. Here's my Get command 
>against a Tab delimted file: (I removed server info and additional fields.. 
>everything else is the same)
>
> http://server:port/appname/solrcore/update/csv?commit=true&header=false&separator=%09&escape=\&stream.file=/location/of/file/on/server/file.csv&fieldnames=id,otherfields
>
>
> My response from solr
>
> <?xml version="1.0" encoding="UTF-8"?>
> <response>
> <lst name="responseHeader"><int name="status">0</int><int 
> name="QTime">591</int></lst>
> </response>
>
> I am experimenting with 2 csv files (1 with 10 records, the other with 1000) 
> to see If I can get this to run correctly before running my entire collection 
> of data. I initially loaded the first 1000 records to an empty core and that 
> seemed to work, however, but when running the above with a csv file that has 
> 10 records, I would like to see only 10 active records in my core. What I get 
> instead, when looking at my stats page:
>
> numDocs 1000
> maxDoc 1010
>
> If I run the same url above while appending an 'optimize=true', I get:
>
> numDocs 1000,
> maxDoc 1000.

A discrepancy between numDocs and maxDoc indicates that there are 
deleted documents in your index.  You might already know this, so here's 
an answer to what I think might be your actual question:

If you want to delete the 1000 existing documents before adding the 10 
documents, then you have to actually do that deletion.  The CSV update 
handler works at a lower level than the DataImport handler, and doesn't 
have "clean" or "full-import" options, which defaults to clean=true. 
The DIH is like a full application embedded inside Solr, one that uses 
an update handler -- it is not itself an update handler.  When 
clean=true or using full-import without a clean option, DIH itself sends 
a "delete all documents" update request.

If you didn't already know the bit about the deleted documents, then 
read this:

It can be normal for indexing "new" documents to cause deleted 
documents.  This happens when you have the same value in your UniqueKey 
field as documents that are already in your index.  Solr knows by the 
config you gave it that they are the same document, so it deletes the 
old one before adding the new one.  Solr has no way to know whether the 
document it already had or the document you are adding is more current, 
so it assumes you know what you are doing and takes care of the deletion 
for you.

When you optimize your index, deleted documents are purged, which is why 
the numbers match there.

Thanks,
Shawn

Reply via email to