Well yes, transformation is required. But it's like data coming from multiple tables.. etc. It's not like getting one row from table and possibly transforming it and using it.
I am thinking perhaps to create some tables (views) that will have the data ready/flattened and then simply feeding it. Cause not sure how much flexibility transformer will give me. Java not number 1 language either :) Thanks for suggestions. Will get a look there. Avlesh Singh wrote: > As Noble has already said, "transforming" content before indexing a very > common requirement. DataImportHandler's Transformer lets you achieve this. > Read up on the same here - > http://wiki.apache.org/solr/DataImportHandler#head-a6916b30b5d7605a990fb03c4ff461b3736496a9 > > Cheers > Avlesh > > 2009/7/22 Noble Paul നോബിള് नोब्ळ् <noble.p...@corp.aol.com> > > >> if each field from the db goes to a separate field in solr as-is . >> Then it is very simple. >> >> if you need to split/join fields before feeding it into solr fields >> you may need to apply transformers >> an example on how your db field looks like and how you wish it to look >> like in solr would be helpful >> >> On Wed, Jul 22, 2009 at 11:57 AM, Julian Davchev<j...@drun.net> wrote: >> >>> Hi folks, >>> >>> I have around 50k documents that are reindexed now and then. >>> >>> Question is what would be the fastest approach to all this. >>> >>> Data is just text ~20fields or so. It comes from database but is first >>> specially formated to get to format suitable for passing in solr. >>> >>> Currently xml post is used but have the feeling this is not optimal for >>> speed wise when it is up to bulk import/reindex. >>> >>> I see http://wiki.apache.org/solr/DataImportHandler but kinda fail to >>> see howto do this specially formated data so solr makes use of it. >>> >>> Are there some real examples,articles on howto use this? >>> >>> >> >> -- >> ----------------------------------------------------- >> Noble Paul | Principal Engineer| AOL | http://aol.com >> >> > >