Hi Mike, I am currently ruling out some bad memory modules. Knowing that this is a index corruption, makes memory corruption more likely. If replacing RAM does not fix the problem (which I need to do anyway due to segmentation faults), I will package up the crash into a reproducible scenario.
On Mon, Aug 18, 2008 at 5:56 AM, Michael McCandless <[EMAIL PROTECTED]> wrote: > > Hi Ian, > > I sent this to java-user, but maybe you didn't see it, so let's try again on > solr-user: > > > It looks like your stored fields file (_X.fdt) is corrupt. > > Are you using multiple threads to add docs? > > Can you try switching to SerialMergeScheduler to verify it's reproducible? > > When you hit this exception, can you stop Solr and then run Lucene's > CheckIndex tool (org.apache.lucene.index.CheckIndex) to verify the > index is corrupt and see which segment it is? Then post back the > exception and "ls -l" of your index directory? > > If you could post the client-side code you're using to build & submit > docs to Solr, and if I can get access to the Medline content, and I > can the repro the bug, then I'll track it down... > > Mike > > On Aug 14, 2008, at 10:18 PM, Ian Connor wrote: > >> I seem to be able to reproduce this very easily and the data is >> medline (so I am sure I can share it if needed with a quick email to >> check). >> >> - I am using fedora: >> %uname -a >> Linux ghetto5.projectlounge.com 2.6.23.1-42.fc8 #1 SMP Tue Oct 30 >> 13:18:33 EDT 2007 x86_64 x86_64 x86_64 GNU/Linux >> %java -version >> java version "1.7.0" >> IcedTea Runtime Environment (build 1.7.0-b21) >> IcedTea 64-Bit Server VM (build 1.7.0-b21, mixed mode) >> - single core (will use shards but each machine just as one HDD so >> didn't see how cores would help but I am new at this) >> - next run I will keep the output to check for earlier errors >> - very and I can share code + data if that will help >> >> On Thu, Aug 14, 2008 at 4:23 PM, Yonik Seeley <[EMAIL PROTECTED]> wrote: >>> >>> Yikes... not good. This shouldn't be due to anything you did wrong >>> Ian... it looks like a lucene bug. >>> >>> Some questions: >>> - what platform are you running on, and what JVM? >>> - are you using multicore? (I fixed some index locking bugs recently) >>> - are there any exceptions in the log before this? >>> - how reproducible is this? >>> >>> -Yonik >>> >>> On Thu, Aug 14, 2008 at 2:47 PM, Ian Connor <[EMAIL PROTECTED]> wrote: >>>> >>>> Hi, >>>> >>>> I have rebuilt my index a few times (it should get up to about 4 >>>> Million but around 1 Million it starts to fall apart). >>>> >>>> Exception in thread "Lucene Merge Thread #0" >>>> org.apache.lucene.index.MergePolicy$MergeException: >>>> java.lang.IndexOutOfBoundsException: Index: 105, Size: 33 >>>> at >>>> org.apache.lucene.index.ConcurrentMergeScheduler.handleMergeException(ConcurrentMergeScheduler.java:323) >>>> at >>>> org.apache.lucene.index.ConcurrentMergeScheduler$MergeThread.run(ConcurrentMergeScheduler.java:300) >>>> Caused by: java.lang.IndexOutOfBoundsException: Index: 105, Size: 33 >>>> at java.util.ArrayList.rangeCheck(ArrayList.java:572) >>>> at java.util.ArrayList.get(ArrayList.java:350) >>>> at >>>> org.apache.lucene.index.FieldInfos.fieldInfo(FieldInfos.java:260) >>>> at org.apache.lucene.index.FieldsReader.doc(FieldsReader.java:188) >>>> at >>>> org.apache.lucene.index.SegmentReader.document(SegmentReader.java:670) >>>> at >>>> org.apache.lucene.index.SegmentMerger.mergeFields(SegmentMerger.java:349) >>>> at >>>> org.apache.lucene.index.SegmentMerger.merge(SegmentMerger.java:134) >>>> at >>>> org.apache.lucene.index.IndexWriter.mergeMiddle(IndexWriter.java:3998) >>>> at org.apache.lucene.index.IndexWriter.merge(IndexWriter.java:3650) >>>> at >>>> org.apache.lucene.index.ConcurrentMergeScheduler.doMerge(ConcurrentMergeScheduler.java:214) >>>> at >>>> org.apache.lucene.index.ConcurrentMergeScheduler$MergeThread.run(ConcurrentMergeScheduler.java:269) >>>> >>>> >>>> When this happens, the disk usage goes right up and the indexing >>>> really starts to slow down. I am using a Solr build from about a week >>>> ago - so my Lucene is at 2.4 according to the war files. >>>> >>>> Has anyone seen this error before? Is it possible to tell which Array >>>> is too large? Would it be an Array I am sending in or another internal >>>> one? >>>> >>>> Regards, >>>> Ian Connor >>>> >>> >> >> >> >> -- >> Regards, >> >> Ian Connor > > -- Regards, Ian Connor