Silly thing … Maybe the immense token was generating because trying to set
"string" as field type for your text ?
Can be ?
Can you wipe out the index, set a proper type for your text, and index
again ?
No worries about the not full stack trace,
We learn and do wrong things everyday :)
Errare humanum est

Cheers

2015-06-19 14:31 GMT+01:00 Paden <rumsey...@gmail.com>:

> Yeah I'm just gonna say hands down this was a totally bad question. My
> fault,
> mea culpa. I'm pretty new to working in an IDE environment and using a
> stack
> trace (I just finished my first year of CS at University and now I'm
> interning). I'm actually kind of embarrassed by how long it took me to
> realize I wasn't looking at the entire stack trace. Idiot moment of the
> week
> for sure. Thanks for the patience guys but when I looked at the entire
> stack
> trace it gave me this.
>
> Caused by: java.lang.IllegalArgumentException: Document contains at least
> one immense term in field="text" (whose UTF8 encoding is longer than the
> max
> length 32766), all of which were skipped.  Please correct the analyzer to
> not produce such terms.  The prefix of the first immense term is: '[84,
> 104,
> 101, 32, 73, 78, 76, 32, 105, 115, 32, 97, 32, 85, 46, 83, 46, 32, 68, 101,
> 112, 97, 114, 116, 109, 101, 110, 116, 32, 111]...', original message:
> bytes
> can be at most 32766 in length; got 44360
>         at
>
> org.apache.lucene.index.DefaultIndexingChain$PerField.invert(DefaultIndexingChain.java:667)
>         at
>
> org.apache.lucene.index.DefaultIndexingChain.processField(DefaultIndexingChain.java:344)
>         at
>
> org.apache.lucene.index.DefaultIndexingChain.processDocument(DefaultIndexingChain.java:300)
>         at
>
> org.apache.lucene.index.DocumentsWriterPerThread.updateDocument(DocumentsWriterPerThread.java:232)
>         at
>
> org.apache.lucene.index.DocumentsWriter.updateDocument(DocumentsWriter.java:458)
>         at
> org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:1350)
>         at
>
> org.apache.solr.update.DirectUpdateHandler2.addDoc0(DirectUpdateHandler2.java:239)
>         at
>
> org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:163)
>         ... 40 more
> Caused by:
> org.apache.lucene.util.BytesRefHash$MaxBytesLengthExceededException: bytes
> can be at most 32766 in length; got 44360
>         at org.apache.lucene.util.BytesRefHash.add(BytesRefHash.java:284)
>         at
> org.apache.lucene.index.TermsHashPerField.add(TermsHashPerField.java:154)
>         at
>
> org.apache.lucene.index.DefaultIndexingChain$PerField.invert(DefaultIndexingChain.java:657)
>         ... 47 more
>
>
> And it took me all of two seconds to realize what had gone wrong. Now I'm
> just trying to figure out how to index the text content without truncating
> all the info or filtering it out entirely, thereby messing up my searching
> capabilities.
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Error-when-submitting-PDF-to-Solr-w-text-fields-using-SolrJ-tp4212704p4212919.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>



-- 
--------------------------

Benedetti Alessandro
Visiting card : http://about.me/alessandro_benedetti

"Tyger, tyger burning bright
In the forests of the night,
What immortal hand or eye
Could frame thy fearful symmetry?"

William Blake - Songs of Experience -1794 England

Reply via email to