Look at the scheme.xml that I provided. I use my own "text_block" type which is derived from "TextField". And I force using StandardTokenizerFactory using tokenizer tag.
If I use StrField type there are no problems with big data indexing. The problem is in the tokenizer. On 23 October 2010 18:55, Yonik Seeley <yo...@lucidimagination.com> wrote: > On Fri, Oct 22, 2010 at 12:07 PM, Sergey Bartunov <sbos....@gmail.com> wrote: >> I'm trying to force solr to index words which length is more than 255 > > If the field is not a text field, the Solr's default analyzer is used, > which currently limits the token to 256 bytes. > Out of curiosity, what's your usecase that you really need a single 34KB > token? > > -Yonik > http://www.lucidimagination.com >