there is another option of passing the table name as a request
parameter and make your sql query templatized .

example
query="select * from ${table}"
and pass the value of table as a request parameter

On Sat, Oct 10, 2009 at 3:52 AM, solr.searcher <solr.searc...@gmail.com> wrote:
>
> Hmmm. Interesting line of thought. Thanks a lot Jay. Will explore this
> approach. There are lot of duplicate tables though :).
>
> I was about to try a different approach - set up two solar cores, keep
> reloading config and updating one, merge with the bigger index ...
>
> But your approach is worth exploring. Thanks.
>
>
>
> Jay Hill wrote:
>>
>> You could use separate DIH config files for each of your three tables.
>> This
>> might be overkill, but it would keep them separate. The DIH is not limited
>> to one request handler setup, so you could create a unique handler for
>> each
>> case with a unique name:
>>
>>    <requestHandler name="/indexer/table1"
>> class="org.apache.solr.handler.dataimport.DataImportHandler">
>>     <lst name="defaults">
>>       <str name="config">table1-config.xml</str>
>>     </lst>
>>   </requestHandler>
>>
>>    <requestHandler name="/indexer/table2"
>> class="org.apache.solr.handler.dataimport.DataImportHandler">
>>      <lst name="defaults">
>>       <str name="config">table2-config.xml</str>
>>     </lst>
>>   </requestHandler>
>>
>>    <requestHandler name="/indexer/table3"
>> class="org.apache.solr.handler.dataimport.DataImportHandler">
>>      <lst name="defaults">
>>       <str name="config">table3-config.xml</str>
>>     </lst>
>>   </requestHandler>
>>
>> When you go to ...solr/admin/dataimport.jsp you should see a list of all
>> DataImportHandlers that are configured, and can select them individually,
>> if
>> that works for your needs.
>>
>> -Jay
>> http://www.lucidimagination.com
>>
>> On Fri, Oct 9, 2009 at 10:57 AM, solr.searcher
>> <solr.searc...@gmail.com>wrote:
>>
>>>
>>> Hi all,
>>>
>>> First of all, please accept my apologies if this has been asked and
>>> answered
>>> before. I tried my best to search and couldn't find anything on this.
>>>
>>> The problem I am trying to solve is as follows. I have multiple tables
>>> with
>>> identical schema - table_a, table_b, table_c ... and I am trying to
>>> create
>>> one big index with the data from each of these tables. The idea was to
>>> programatically create the data-config file (just changing the table
>>> name)
>>> and do a reload-config followed by a full-import with clean set to false.
>>> In
>>> other words:
>>>
>>> 1. publish the data-config file
>>> 2. do a reload-config
>>> 3. do a full-import with clean = false
>>> 4. commit, optimize
>>> 5. repeat with new table name
>>>
>>> I wanted to then follow the same procedure for delta imports. The problem
>>> is
>>> that after I do a reload-config and then do a full-import, the old data
>>> in
>>> the index is getting lost.
>>>
>>> What am I missing here? Please note that I am new to solr.
>>>
>>> INFO: [] webapp=/solr path=/dataimport
>>> params={command=reload-config&clean=false} status=0 QTime=4
>>> Oct 9, 2009 10:17:30 AM org.apache.solr.core.SolrCore execute
>>> INFO: [] webapp=/solr path=/dataimport
>>> params={command=full-import&clean=false} status=0 QTime=1
>>> Oct 9, 2009 10:17:30 AM org.apache.solr.handler.dataimport.SolrWriter
>>> readIndexerProperties
>>> INFO: Read dataimport.properties
>>> Oct 9, 2009 10:17:30 AM org.apache.solr.handler.dataimport.DataImporter
>>> doFullImport
>>> INFO: Starting Full Import
>>> Oct 9, 2009 10:17:30 AM org.apache.solr.handler.dataimport.SolrWriter
>>> readIndexerProperties
>>> INFO: Read dataimport.properties
>>> Oct 9, 2009 10:17:30 AM
>>> org.apache.solr.handler.dataimport.JdbcDataSource$1
>>> call
>>> INFO: Creating a connection for entity <blah blah blah>
>>> Oct 9, 2009 10:17:30 AM
>>> org.apache.solr.handler.dataimport.JdbcDataSource$1
>>> call
>>> INFO: Time taken for getConnection(): 12
>>> Oct 9, 2009 10:17:31 AM org.apache.solr.core.SolrDeletionPolicy onInit
>>> INFO: SolrDeletionPolicy.onInit: commits:num=1
>>>
>>>
>>> commit{dir=/blah/blah/index,segFN=segments_1z,version=1255032607825,generation=71,filenames=[segments_1z,
>>> _cl.cfs]
>>> Oct 9, 2009 10:17:31 AM org.apache.solr.core.SolrDeletionPolicy
>>> updateCommits
>>> INFO: last commit = 1255032607825
>>>
>>> Any help will be greatly appreciated. Is there any other way to
>>> automaticaly
>>> slurp data from multiple, identical tables?
>>>
>>> Thanks a lot.
>>>
>>> --
>>> View this message in context:
>>> http://www.nabble.com/Dynamic-Data-Import-from-multiple-identical-tables-tp25825381p25825381.html
>>> Sent from the Solr - User mailing list archive at Nabble.com.
>>>
>>>
>>
>>
>
> --
> View this message in context: 
> http://www.nabble.com/Dynamic-Data-Import-from-multiple-identical-tables-tp25825381p25828773.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>



-- 
-----------------------------------------------------
Noble Paul | Principal Engineer| AOL | http://aol.com

Reply via email to