In an older wiki post (which I know is outdated...but still...) located here: 
https://wiki.apache.org/solr/FileBasedSpellChecker, the end of the first 
paragraph  (the intro) indicates that "...it isn't all that hard to create an 
index from a file and have it weight the terms."

I've been thinking about this, and it would be PERFECT for what we need 
(because we HAVE a dictionary from a different search engine WITH weights on it 
that have been crafted/tweaked over a long period of time). However, as I've 
read the spellcheck documentation, I'm really not certain what this person 
(Mark Bennett) was talking about.

Question #1: Has anyone pursued this possibility? If so, can you offer some 
insight as to how you approached this?

Question #2: I'd be grateful for someone's "hint" as to where to start looking 
to try to create this workaround if you have some experience with it. I'm 
guessing that the only way to do this is to dive into the File based spell 
check dictionary code and then clone/modify what is there. I was just wondering 
if I was missing something or if someone was aware of a quicker way to scale 
this technological hill. I realize I could dissect the format of the spellcheck 
dictionary and load it as I desire...but I'm a bit concerned about this as this 
is for a long-lived project that is expected to live a lot longer, and this 
approach seems to be a bit "brittle." It will break the first time Apache makes 
the slightest change to the structure of the spell check index.

If anyone has any insights to offer I'd appreciate it.

Thanks.

Peter S. Lee

Reply via email to