Look at the scheme.xml that I provided. I use my own "text_block" type
which is derived from "TextField". And I force using
StandardTokenizerFactory using tokenizer tag.

If I use StrField type there are no problems with big data indexing.
The problem is in the tokenizer.

On 23 October 2010 18:55, Yonik Seeley <yo...@lucidimagination.com> wrote:
> On Fri, Oct 22, 2010 at 12:07 PM, Sergey Bartunov <sbos....@gmail.com> wrote:
>> I'm trying to force solr to index words which length is more than 255
>
> If the field is not a text field, the Solr's default analyzer is used,
> which currently limits the token to 256 bytes.
> Out of curiosity, what's your usecase that you really need a single 34KB 
> token?
>
> -Yonik
> http://www.lucidimagination.com
>

Reply via email to