I think you should replace your new lucene-core-2.9.3-dev.jar in \apache-solr-1.4.1\lib and then create a new solr.war under \apache-solr-1.4.1\dist. And copy this new solr.war to solr/example/webapps/solr.war
--- On Sat, 10/23/10, Sergey Bartunov <sbos....@gmail.com> wrote: > From: Sergey Bartunov <sbos....@gmail.com> > Subject: Re: How to index long words with StandardTokenizerFactory? > To: solr-user@lucene.apache.org > Date: Saturday, October 23, 2010, 5:45 PM > Yes. I did. Won't help. > > On 23 October 2010 17:45, Ahmet Arslan <iori...@yahoo.com> > wrote: > > Did you delete the folder > Jetty_0_0_0_0_8983_solr.war_** under > apache-solr-1.4.1\example\work? > > > > --- On Sat, 10/23/10, Sergey Bartunov <sbos....@gmail.com> > wrote: > > > >> From: Sergey Bartunov <sbos....@gmail.com> > >> Subject: Re: How to index long words with > StandardTokenizerFactory? > >> To: solr-user@lucene.apache.org > >> Date: Saturday, October 23, 2010, 3:56 PM > >> Here are all the files: http://rghost.net/3016862 > >> > >> 1) StandardAnalyzer.java, StandardTokenizer.java - > patched > >> files from > >> lucene-2.9.3 > >> 2) I patch these files and build lucene by typing > "ant" > >> 3) I replace lucene-core-2.9.3.jar in solr/lib/ by > my > >> lucene-core-2.9.3-dev.jar that I'd just compiled > >> 4) than I do "ant compile" and "ant dist" in solr > folder > >> 5) after that I recompile > solr/example/webapps/solr.war > >> with my new > >> solr and lucene-core jars > >> 6) I put my schema.xml in solr/example/solr/conf/ > >> 7) then I do "java -jar start.jar" in > solr/example > >> 8) index big_post.xml > >> 9) trying to find this document by "curl > >> http://localhost:8983/solr/select?q=body:big*" > >> (big_post.xml contains > >> a long word bigaaaaa...aaaa) > >> 10) solr returns nothing > >> > >> On 23 October 2010 02:43, Steven A Rowe <sar...@syr.edu> > >> wrote: > >> > Hi Sergey, > >> > > >> > What does your ~34kb field value look like? > Does > >> StandardTokenizer think it's just one token? > >> > > >> > What doesn't work? What happens? > >> > > >> > Steve > >> > > >> >> -----Original Message----- > >> >> From: Sergey Bartunov [mailto:sbos....@gmail.com] > >> >> Sent: Friday, October 22, 2010 3:18 PM > >> >> To: solr-user@lucene.apache.org > >> >> Subject: Re: How to index long words > with > >> StandardTokenizerFactory? > >> >> > >> >> I'm using Solr 1.4.1. Now I'm successed > with > >> replacing lucene-core jar > >> >> but maxTokenValue seems to be used in > very strange > >> way. Currenty for > >> >> me it's set to 1024*1024, but I couldn't > index a > >> field with just size > >> >> of ~34kb. I understand that it's a little > weird to > >> index such a big > >> >> data, but I just want to know it doesn't > work > >> >> > >> >> On 22 October 2010 20:36, Steven A Rowe > <sar...@syr.edu> > >> wrote: > >> >> > Hi Sergey, > >> >> > > >> >> > I've opened an issue to add a > maxTokenLength > >> param to the > >> >> StandardTokenizerFactory configuration: > >> >> > > >> >> > https://issues.apache.org/jira/browse/SOLR-2188 > >> >> > > >> >> > I'll work on it this weekend. > >> >> > > >> >> > Are you using Solr 1.4.1? I ask > because of > >> your mention of Lucene > >> >> 2.9.3. I'm not sure there will ever be > a Solr > >> 1.4.2 release. I plan on > >> >> targeting Solr 3.1 and 4.0 for the > SOLR-2188 fix. > >> >> > > >> >> > I'm not sure why you didn't get the > results > >> you wanted with your Lucene > >> >> hack - is it possible you have other > Lucene jars > >> in your Solr classpath? > >> >> > > >> >> > Steve > >> >> > > >> >> >> -----Original Message----- > >> >> >> From: Sergey Bartunov [mailto:sbos....@gmail.com] > >> >> >> Sent: Friday, October 22, 2010 > 12:08 PM > >> >> >> To: solr-user@lucene.apache.org > >> >> >> Subject: How to index long words > with > >> StandardTokenizerFactory? > >> >> >> > >> >> >> I'm trying to force solr to > index words > >> which length is more than 255 > >> >> >> symbols (this constant is > >> DEFAULT_MAX_TOKEN_LENGTH in lucene > >> >> >> StandardAnalyzer.java) using > >> StandardTokenizerFactory as 'filter' tag > >> >> >> in schema configuration XML. > Specifying > >> the maxTokenLength attribute > >> >> >> won't work. > >> >> >> > >> >> >> I'd tried to make the dirty > hack: I > >> downloaded lucene-core-2.9.3 src > >> >> >> and changed the > DEFAULT_MAX_TOKEN_LENGTH > >> to 1000000, built it to jar > >> >> >> and replaced original > lucene-core jar in > >> solr /lib. But seems like > >> >> >> that it had bring no effect. > >> > > > > > > > > >