> 
> Hello all,
> 
> Greetings!!,
> 
> Environment: 
> I have been working on a solr instance with  requirement of ~ 40 columns and 
> 50000 records of data.i am using data import handler to configure the data 
> load and indexing..also I use regex transformer to read the data file(csv) 
> and apply regex to parse individual columns.note that all the columns are set 
> with index option as true .configuration is given below
> 
> Issue:
> While loading data with 15 columns and 50000 records hardly takes around 2 
> mins, increasing the number of columns increase the data load time very 
> high..for e.g) 24 columns with 50000 records takes nearly 30 mins.i don't see 
> any increased memory consumption with solr admin dashboard .on the other 
> hand,using the update csv handler takes hardly around one min to load the 
> complete data
> 
> Please provide Any suggestions/alternatives to improve the performance 
> (.please note I couldn't use csv handler since with csv handler, there's not  
>  an query able option to know the status and the need to delete the index 
> every time before data load )
> 
> Configuration details
> 
> Solr index size 14 MB( when used update csv handler)
> Data file size 5 MB
> 
> 
> Data- config.xml
> 
> Datasource type - Filedatasource    encoding UTF8 
> 
> Transformer - 
> regextransformer
> 
> 
> Regex ="^(.*),(.*)............"
> 
> groupNames = "field1,field2,......field40"
> 
> 
> 
> Regards 
> Ibrahim
> 
> 

Reply via email to