neosky wrote
> 
> I use the solr 3.5 version
> 1. It seems that the Ngramtokenizerfactory only token the first 1024
> characters. I search the problem on the Internet, somebody had noticed the
> bug in 2007, but I can't find the solution.
> ps:  my max field length has been modified 
> <maxFieldLength>50000</maxFieldLength>
> This is very critical for me.
> 
> It is not fixed as I know. In the NGramTokenizer
>  char[] chars = new char[1024];
>       input.read(chars);
> but I don't know what's the different between NGramTokenizer and
> NGramTokenFilter
> suppose I want to write my Analyzier which should I use?
> 
> 
> 2.the second questions that when I defines the 
>  minGramSize=3
>  maxGramSize=8
>  what happens when I search a query length is 5. Does it work?
> My consideration is to use the copyfiled to specify the gram from 3,8, I
> am not sure it is a solution.I am very worry about the index speed. I
> spend more than 6 hours to index the gram from 7,8 for testing.
> Thanks!
> 

I still need time to index to test.

--
View this message in context: 
http://lucene.472066.n3.nabble.com/Two-questions-about-the-Ngramtokenizerfactory-tp3893045p3894851.html
Sent from the Solr - User mailing list archive at Nabble.com.

Reply via email to