> Dallas, got money to spend on solving this problem? I > believe this is something that tools like LingPipe can solve > through language model training and named entity extraction.
Hi Otis, Thank-you for your reply. I'm familiar with tools like LingPipe, but this problem is actually *much* simpler. The places have already been entered into Place fields. (I don't have to try to identify place names in running text, which is what tools like LingPipe are for.) All I have to do is convert user-entered place names that may omit levels or contain abbreviations like "Chicago, IL", into complete place names like "Chicago, Cook, Illinois, United States". I have an algorithm already written to do this; I just don't know how to call it from a token filter because I don't know how to access SolrCore from within a token filter object. If I cannot (or for some reason should not) access SolrCore from a token filter, my alternative is before indexing a document, to issue a query to convert the place fields associated with that document into complete place names, then pass in the document with the complete place names into SOLR for indexing. By issuing this query instead from the token filter if possible, I was hoping to avoid the extra round-trip query-response between the indexing process and the SOLR server that this would entail. Thanks again! -dallan