Wow - great stuff Steve!
As for StandardTokenizer and Java version - no worries there really,
as Solr itself requires Java 1.5+, so when such a tokenizer is made
available it could be used just fine in Solr even if it isn't built
into a core Lucene release for a while.
Erik
On Feb 28, 2008, at 12:08 PM, Steven A Rowe wrote:
On 02/28/2008 at 11:26 AM, Ken Krugler wrote:
And as Erik mentioned, it appears that line 114 of
StandardTokenizerImpl.jflex:
http://www.krugle.org/kse/files/svn/svn.apache.org/lucene/java/
trunk/src/java/org/apache/lucene/analysis/standard/
StandardTokenizerImpl.jflex
needs to be updated to include the Extension B character range.
JFlex 1.4.1 (the latest release) does not support supplementary
code points (those above the BMP - Basic Multilingual Plane: [U
+0000-U+FFFF]), and CJK Ideograph Extension B is definitely a
supplementary range - see the first column from <http://
www.unicode.org/Public/3.1-Update/UnicodeData-3.1.0.txt> (the
extent of this range is unchanged through the latest [beta]
version, 5.1.0):
20000;<CJK Ideograph Extension B, First> ...
2A6D6;<CJK Ideograph Extension B, Last> ...
I am working with Gerwin Klein on the development version of JFlex,
and am hoping to get Level 1 [Regular Expression] Basic Unicode
Support into the next release (see <http://unicode.org/reports/tr18/
>) - among other things, this entails accepting supplementary code
points.
However, the next release of JFlex will require Java 1.5+, and
Lucene 2.X requires Java 1.4, so until Lucene reaches release 3.0
and begins requiring Java 1.5 (and Solr incorporates it), JFlex
support of supplementary code points is moot.
In short, it'll probably be at least a year before the
StandardTokenizer can be modified to accept supplementary
characters, given the processes involved.
Steve