Looking at it, and now knowing how much memory your other processes on your box 
use (nor how much memory you have set aside for Java), I would start with 
DOUBLING your ram. Make sure that you have enough Java memory.

You will know if it has some effect by using the 2:1 size ratio. 100mb for all 
that data ia pretty small, I think.


Use the scientific method; Change only one parameter at a time and check 
results.

It's always on of four things:
(in different order depending on task, but listed alphabetically here)
------------------------------
Memory (process assigned and/or actual physical memory)
Processor
Network Bandwidth
Hard Drive Bandwidth
(sometimes you can add motherboard I/O paths also.
 as of this date, AMD has much more I/O paths in their
 consumer line of processors.)

In order ease of experimenting with(Easiest to hardest):
-----------------------------------
Appication/process assigned memory
Physical memory
Network Bandwidth
HardDrive Bandwidth
  Screaming fast SCSI 15K rpm drives
  RAID arrays, casual
  RAID arrays, professional
  External DRAM drive 64 gig max/RAID them for more
Processor(s) 
  Put maximum speed/cache size motherboard will take.
  Otherwise, USUALLY requires changing motherboard/HOSTING setup
I/O channels
  USUALLY requires changing motherboard/HOSTING setup





Dennis Gearon

Signature Warning
----------------
It is always a good idea to learn from your own mistakes. It is usually a 
better idea to learn from others’ mistakes, so you do not have to make them 
yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'

EARTH has a Right To Life,
  otherwise we all die.


--- On Sat, 10/9/10, sivaprasad <sivaprasa...@echidnainc.com> wrote:

> From: sivaprasad <sivaprasa...@echidnainc.com>
> Subject: Re: Speeding up solr indexing
> To: solr-user@lucene.apache.org
> Date: Saturday, October 9, 2010, 8:09 AM
> 
> Hi,
> Please find the configurations below.
> 
> Machine configurations(Solr running here):
> 
> RAM - 4 GB
> HardDisk - 180GB
> Os - Red Hat linux version 5
> Processor-2x Intel Core 2 Duo CPU @2.66GHz
> 
> 
> 
> Machine configurations(Mysql server is running here):
> RAM - 4 GB
> HardDisk - 180GB
> Os - Red Hat linux version 5
> Processor-2x Intel Core 2 Duo CPU @2.66GHz
> 
> My sql Server deatils:
> My sql version - Mysql 5.0.22
> 
> Solr configuration details:
> 
>  <indexDefaults>
>   
>    
> <useCompoundFile>false</useCompoundFile>
> 
>     <mergeFactor>20</mergeFactor>
>    
>    
> <!--<maxBufferedDocs>1000</maxBufferedDocs>--> 
>   
>    
> <ramBufferSizeMB>100</ramBufferSizeMB>
>    
> <maxMergeDocs>2147483647</maxMergeDocs>
>    
> <maxFieldLength>10000</maxFieldLength>
>    
> <writeLockTimeout>1000</writeLockTimeout>
>    
> <commitLockTimeout>10000</commitLockTimeout>
>    
> <!--<luceneAutoCommit>false</luceneAutoCommit>-->
>    
>    
> <!--<mergePolicy>org.apache.lucene.index.LogByteSizeMergePolicy</mergePolicy>-->
>     
>    
> <!--<mergeScheduler>org.apache.lucene.index.ConcurrentMergeScheduler</mergeScheduler>-->
>     <lockType>single</lockType>
>   </indexDefaults>
> 
>   <mainIndex>
>     
>    
> <useCompoundFile>false</useCompoundFile>
>    
> <ramBufferSizeMB>100</ramBufferSizeMB>
>     <mergeFactor>20</mergeFactor>
>    
>    
> <!--<maxBufferedDocs>1000</maxBufferedDocs>-->
>    
> <maxMergeDocs>2147483647</maxMergeDocs>
>    
> <maxFieldLength>10000</maxFieldLength>
>    
> <unlockOnStartup>false</unlockOnStartup>
>   </mainIndex>
> 
>   <!-- the default high-performance update handler
> -->
>   <updateHandler
> class="solr.DirectUpdateHandler2">
>    
> <maxPendingDeletes>100000</maxPendingDeletes>
>     <autoCommit> 
>       <maxDocs>10000</maxDocs> 
>       <maxTime>60000</maxTime>
>     </autoCommit>
>     
>     <!-- A postCommit event is fired after
> every commit or optimize command
>     <listener event="postCommit"
> class="solr.RunExecutableListener">
>       <str
> name="exe">solr/bin/snapshooter</str>
>       <str name="dir">.</str>
>       <bool
> name="wait">true</bool>
>       <arr name="args">
> <str>arg1</str> <str>arg2</str>
> </arr>
>       <arr name="env">
> <str>MYVAR=val1</str> </arr>
>     </listener>
>     -->
>     <!-- A postOptimize event is fired only
> after every optimize command,
> useful
>          in conjunction with
> index distribution to only distribute optimized
> indicies 
>     <listener event="postOptimize"
> class="solr.RunExecutableListener">
>       <str
> name="exe">snapshooter</str>
>       <str
> name="dir">solr/bin</str>
>       <bool
> name="wait">true</bool>
>     </listener>
>     -->
>   </updateHandler>
> 
> Solr document details:
> 
> 21 fields are indexed and stored
> 3 fileds are indexed only.
> 3 fileds are stored only.
> 3 fileds are indexed,stored and multi valued
> 2 fileds indexed and multi valued
> 
> And i am copying some of the indexed fileds.In this 2
> fileds are multivalued
> and has thousands of values.
> 
> In db-config-file the main table contains 0.6 million
> records.
> 
> When i tested for the same records, the index has taken 1hr
> 30 min.In this
> case one of the multivalued filed table doesn't have
> records.After putting
> data into this table,for each main table record , this
> table has thousands
> of records and this filed is indexed and stored.It is
> taking more than 24
> hrs .
> 
> Solr is running on tomcat 6.0.26, jdk1.6.0_17 and solr
> 1.4.1
> 
> I am using JVM's default settings.
> 
> Why this is taking this much time?Any body has suggestions,
> where i am going
> wrong.
> 
> Thanks,
> JS
> -- 
> View this message in context: 
> http://lucene.472066.n3.nabble.com/Speeding-up-solr-indexing-tp1667054p1670737.html
> Sent from the Solr - User mailing list archive at
> Nabble.com.
>

Reply via email to