Re: Control Solr spellcheck functionality to provide suggestions for correct word
Hi Rohan, Sharing the configuration: text_suggest default title solr.DirectSolrSpellChecker internal 0.5 2 1 5 4 0.01 Also, below are the parameters, passing into Solr params: solrParams.add("spellcheck.dictionary", "default"); solrParams.add("spellcheck.extendedResults", "true"); solrParams.add("spellcheck.count", "10"); solrParams.add("spellcheck.alternativeTermCount", "5"); solrParams.add("spellcheck.maxResultsForSuggest", "5"); solrParams.add("spellcheck.collate", "true"); solrParams.add("spellcheck.collateExtendedResults", "true"); solrParams.add("spellcheck.maxCollationTries", "10"); solrParams.add("spellcheck.maxCollations", "5"); solrParams.add("spellcheck.onlyMorePopular", "false"); is Solr always provide the suggestions for correct words also ? Any pointers. -- Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
How to list all collections in solr-4.7.2
Hi all, I have setup two solr-4.7.2 server instances on two diff machines with 3 zookeeper severs in solrcloud mode. Now, I want to retrieve list of all the collections that I have created in solrcloud mode. I tried LIST command of collections api, but its not working with solr-4.7.2. Error: unknown command LIST Please suggest me the command, that I can use. Thanks.
Indexing using SolrOutputFormat class
Hi All, I am using Solr-5.0.0 API for indexing data in our application and the requirement is to index the data in batches, using solr-mapreduce API. In our application, we may receive data from any type of input source for example: file, streams and any other relational or non-relational Db’s in a particular format. And I need to index this data into Solr, by using SolrOutputFormat class. As per my analysis until now, I find that SolrOutputFormat works with the EmbeddedSolrServer and requires path to config files for indexing data, without the need of passing host and port for creating the SolrClient. I checked for the documentation online, but couldn’t find any proper examples that make the use of SolrOutputFormat class. Does anybody have some implementations or a document, which mentions details like what exactly needs to be passed as input to SolrOutputFormat configuration, etc.? Any pointers would be helpful.
Fwd: Question to send
HI, I have one running solr core with some data indexed on solr server. This core is designed to provide OpenNLP functionalities for indexing and searching. So I have kept following binary models at this location: *\apache-tomcat-7.0.53\solr\collection1\conf\opennlp * · en-sent.bin · en-token.bin · en-pos-maxent.bin · en-ner-person.bin · en-ner-location.bin *My Problem is*: When I unload the running core, and try to delete conf directory from it. It is not allowing me to delete directory with prompt that *en-sent.bin*and *en-token.bin* is in use. If I have unloaded core, then why it is not unlocking the connection with core? Is this a known issue with OpenNLP Binaries? How can I release the connection between unloaded core and conf directory. (Specially binary models) Please provide me some pointers on this. Thanks in Advance
SOLR Performance Benchmarking
Hi, I am using SolrMeter for performance benchmarking. I am able to successfully test my solr setup up to 1000 queries per min while searching. But when I am exceeding this limit say 1500 search queries per min, facing "Server Refused Connection" in SOLR. Currently, I have only one solr server running on 64-bit 4 GB ram machine for testing. Please provide me some pointers , to optimize SOLR so that it can handle large number of request. (Specially more than 1000 request per min). Is there any change that I can do in solrconfig.xml or some other change to support this? Thanks in Advance DISCLAIMER == This e-mail may contain privileged and confidential information which is the property of Persistent Systems Ltd. It is intended only for the use of the individual or entity to which it is addressed. If you are not the intended recipient, you are not authorized to read, retain, copy, print, distribute or use this message. If you have received this communication in error, please notify the sender and delete all copies of this message. Persistent Systems Ltd. does not accept any liability for virus infected mails.
SOLR Performance benchmarking
Hi, I am using SolrMeter for load/stress testing solr performance. Tomcat is configured with default "maxThreads" (i.e. 200). I set Intended Request per min in SolrMeter to 1500 and performed testing. I found that sometimes it works with this much load on solr but sometimes it gives error "Sever Refused Connection" in solr. On getting this error, i increased maxThreads to some higher value, and then it works again. I would like to know why solr is behaving abnormally, initially when it was working with maxThreads=200. Please provide me some pointers.
Need Help for Location searching
Hi, I wanted to design an analyzer that can support location containment relationship For example Europe->France->Paris My requirement is like: when a user search for any country , then results must have the documents having that country , as well as the documents having states and cities which comes under that country. But, documents with country name must have high relevancy. And the same when a user search for state or city. It must obeys containment relationship up to 4 levels .i.e. Continent->Country->State->City Also, I have designed analyzer using synonym filter factory for the same, and its working as per expectation. But I wanted to know, is there any another way, apart from using synonym filter factory that can be used for the same. Is SOLR provide any tokenziers or filters for this? Please provide me some pointers to move ahead. Thanks in Advance
Re: need help on OpenNLP with Solr
Hi, Also i wanted know, Is it possible to integrate wordnet with this analyzer? I want to use wordnet as synonym expansion along with OpenNLP filters. What are the changes required in solr schema.xml and solrconfig.xml? Thanks in Advance On Mon, Jan 6, 2014 at 9:37 PM, rashi gandhi wrote: > Hi, > > > > I have applied OpenNLP (LUCENE 2899.patch) patch to SOLR-4.5.1 for nlp > searching and it is working fine. > > Also I have designed an analyzer for this: > > positionIncrementGap="100"> > > > > sentenceModel="opennlp/en-test-sent.bin" >tokenizerModel="opennlp/en-test-tokenizer.bin"/> > > ignoreCase="true" words="stopwords.txt" enablePositionIncrements="true"/> > > posTaggerModel="opennlp/en-pos-maxent.bin"/> > > nerTaggerModels="opennlp/en-ner-person.bin"/> > > nerTaggerModels="opennlp/en-ner-location.bin"/> > > class="solr.LowerCaseFilterFactory"/> > > class="solr.SnowballPorterFilterFactory"/> > > > > > > sentenceModel="opennlp/en-test-sent.bin" tokenizerModel > ="opennlp/en-test-tokenizer.bin"/> > > ignoreCase="true" words="stopwords.txt" enablePositionIncrements="true"/> > > posTaggerModel="opennlp/en-pos-maxent.bin"/> > > nerTaggerModels="opennlp/en-ner-person.bin"/> > > nerTaggerModels="opennlp/en-ner-location.bin"/> > > class="solr.LowerCaseFilterFactory"/> > > class="solr.SnowballPorterFilterFactory"/> > > > > > > > I am able to find that posTaggerModel is performing tagging in the phrases > and add the payloads. ( but iam not able to analyze it) > > My Question is: > Can i search a phrase giving high boost to NOUN then VERB ? > For example: if iam searching "sitting on blanket" , so i want to give > high boost to NOUN term first then VERB, that are tagged by OpenNLP. > How can i use payloads for boosting? > What are the changes required in schema.xml? > > Please provide me some pointers to move ahead > > Thanks in advance > > > >
need help on OpenNLP with Solr
Hi, I have applied OpenNLP (LUCENE 2899.patch) patch to SOLR-4.5.1 for nlp searching and it is working fine. Also I have designed an analyzer for this: I am able to find that posTaggerModel is performing tagging in the phrases and add the payloads. ( but iam not able to analyze it) My Question is: Can i search a phrase giving high boost to NOUN then VERB ? For example: if iam searching "sitting on blanket" , so i want to give high boost to NOUN term first then VERB, that are tagged by OpenNLP. How can i use payloads for boosting? What are the changes required in schema.xml? Please provide me some pointers to move ahead Thanks in advance
Re: need help on OpenNLP with Solr
Thanks Lance for clearing. One more Question: Is there a possibility of integrating boosting with payloads in LUCENE-2899 with solr? Thanks in Advance On Mon, Jan 6, 2014 at 9:46 PM, rashi gandhi wrote: > Hi, > > Also i wanted know, > Is it possible to integrate wordnet with this analyzer? > I want to use wordnet as synonym expansion along with OpenNLP filters. > What are the changes required in solr schema.xml and solrconfig.xml? > > Thanks in Advance > > > On Mon, Jan 6, 2014 at 9:37 PM, rashi gandhi wrote: > >> Hi, >> >> >> >> I have applied OpenNLP (LUCENE 2899.patch) patch to SOLR-4.5.1 for nlp >> searching and it is working fine. >> >> Also I have designed an analyzer for this: >> >> > positionIncrementGap="100"> >> >> >> >> > sentenceModel="opennlp/en-test-sent.bin" >>tokenizerModel="opennlp/en-test-tokenizer.bin"/> >> >> > ignoreCase="true" words="stopwords.txt" enablePositionIncrements="true"/> >> >> > posTaggerModel="opennlp/en-pos-maxent.bin"/> >> >> > nerTaggerModels="opennlp/en-ner-person.bin"/> >> >> > nerTaggerModels="opennlp/en-ner-location.bin"/> >> >> > class="solr.LowerCaseFilterFactory"/> >> >> > class="solr.SnowballPorterFilterFactory"/> >> >> >> >> >> >> > sentenceModel="opennlp/en-test-sent.bin" tokenizerModel >> ="opennlp/en-test-tokenizer.bin"/> >> >> > ignoreCase="true" words="stopwords.txt" enablePositionIncrements="true"/> >> >> > posTaggerModel="opennlp/en-pos-maxent.bin"/> >> >> > nerTaggerModels="opennlp/en-ner-person.bin"/> >> >> > nerTaggerModels="opennlp/en-ner-location.bin"/> >> >> > class="solr.LowerCaseFilterFactory"/> >> >> > class="solr.SnowballPorterFilterFactory"/> >> >> >> >> >> >> >> I am able to find that posTaggerModel is performing tagging in the >> phrases and add the payloads. ( but iam not able to analyze it) >> >> My Question is: >> Can i search a phrase giving high boost to NOUN then VERB ? >> For example: if iam searching "sitting on blanket" , so i want to give >> high boost to NOUN term first then VERB, that are tagged by OpenNLP. >> How can i use payloads for boosting? >> What are the changes required in schema.xml? >> >> Please provide me some pointers to move ahead >> >> Thanks in advance >> >> >> >> >
Fwd: Need help for integrating solr-4.5.1 with UIMA
r.run(JIoEndpoint.java:310) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:619) Caused by: java.net.ConnectException: Connection timed out: connect at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333) at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195) at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366) at java.net.Socket.connect(Socket.java:529) at java.net.Socket.connect(Socket.java:478) at sun.net.NetworkClient.doConnect(NetworkClient.java:163) at sun.net.www.http.HttpClient.openServer(HttpClient.java:394) at sun.net.www.http.HttpClient.openServer(HttpClient.java:529) at sun.net.www.http.HttpClient.(HttpClient.java:233) at sun.net.www.http.HttpClient.New(HttpClient.java:306) at sun.net.www.http.HttpClient.New(HttpClient.java:323) at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:860) at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:801) at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:726) at sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:904) at org.apache.uima.annotator.calais.OpenCalaisAnnotator.callServiceOnText(OpenCalaisAnnotator.java:219) at org.apache.uima.annotator.calais.OpenCalaisAnnotator.process(OpenCalaisAnnotator.java:126) ... 35 more What is going wrong? Please help me on this. Thanks in Advance Rashi
Re: Need help for integrating solr-4.5.1 with UIMA
:953) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:408) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1008) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:589) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:310) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:619) Caused by: org.apache.uima.resource.ResourceInitializationException at org.apache.lucene.analysis.uima.ae.BasicAEProvider.getAE(BasicAEProvider.java:58) at org.apache.solr.uima.processor.UIMAUpdateRequestProcessorFactory.getInstance(UIMAUpdateRequestProcessorFactory.java:61) ... 22 more Caused by: java.lang.NullPointerException at org.apache.uima.util.XMLInputSource.(XMLInputSource.java:118) at org.apache.lucene.analysis.uima.ae.BasicAEProvider.getInputSource(BasicAEProvider.java:84) at org.apache.lucene.analysis.uima.ae.BasicAEProvider.getAE(BasicAEProvider.java:50) ... 23 more i think solr is not able to load the descriptor file iam defining in tab. Please provide me some help on this. Thanks in Advance Rashi On Mon, Feb 3, 2014 at 2:50 PM, rashi gandhi wrote: > Hi, > > > > I'm trying to integrate Solr 4.5.1 with UIMA and following the steps of > the solr-4.5.1\contrib\uima\readme.txt. > > Edited the solrconfig.xml as given in readme.txt. Also I have registered > the required keys. > > > > But each time when I am indexing data , solr returns error: > > > > Feb 3, 2014 2:04:32 PM > org.apache.uima.analysis_engine.impl.PrimitiveAnalysisEngine_impl > callAnalysisComponentProcess(405) > > SEVERE: Exception occurred > > org.apache.uima.analysis_engine.AnalysisEngineProcessException > > at > org.apache.uima.annotator.calais.OpenCalaisAnnotator.process(OpenCalaisAnnotator.java:206) > > at > org.apache.uima.analysis_component.CasAnnotator_ImplBase.process(CasAnnotator_ImplBase.java:56) > > at > org.apache.uima.analysis_engine.impl.PrimitiveAnalysisEngine_impl.callAnalysisComponentProcess(PrimitiveAnalysisEngine_impl.java:377) > > at > org.apache.uima.analysis_engine.impl.PrimitiveAnalysisEngine_impl.processAndOutputNewCASes(PrimitiveAnalysisEngine_impl.java:295) > > at > org.apache.uima.analysis_engine.asb.impl.ASB_impl$AggregateCasIterator.processUntilNextOutputCas(ASB_impl.java:567) > > at > org.apache.uima.analysis_engine.asb.impl.ASB_impl$AggregateCasIterator.(ASB_impl.java:409) > > at > org.apache.uima.analysis_engine.asb.impl.ASB_impl.process(ASB_impl.java:342) > > at > org.apache.uima.analysis_engine.impl.AggregateAnalysisEngine_impl.processAndOutputNewCASes(AggregateAnalysisEngine_impl.java:267) > > at > org.apache.uima.analysis_engine.impl.AnalysisEngineImplBase.process(AnalysisEngineImplBase.java:267) > > at > org.apache.uima.analysis_engine.impl.AnalysisEngineImplBase.process(AnalysisEngineImplBase.java:280) > > at > org.apache.solr.uima.processor.UIMAUpdateRequestProcessor.processText(UIMAUpdateRequestProcessor.java:173) > > at > org.apache.solr.uima.processor.UIMAUpdateRequestProcessor.processAdd(UIMAUpdateRequestProcessor.java:79) > > at > org.apache.solr.handler.loader.XMLLoader.processUpdate(XMLLoader.java:247) > > at > org.apache.solr.handler.loader.XMLLoader.load(XMLLoader.java:174) > > at > org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:92) > > at > org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:74) > > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) > > at > org.apache.solr.core.SolrCore.execute(SolrCore.java:1859) > > at > org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:703) > > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:406) > > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:195) > > at > org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Applica
Fwd: help on edismax_dynamic fields
Hello, I am using edismax parser in my project. I just wanted to confirm whether we can use dynamic fields with edismax or not. When I am using specific dynamic field in qf or pf parameter , it is working. But when iam using dynamic fields with *, like this: explicit 10 text edismax * * * *_nlp_new_sv^0.8* * *_nlp_copy_sv^0.2* * * It is not working. Is it possible to use dynamic fields with *, like mentioned above with edismax? Please provide me some pointers on this. Thanks in advance.
Fwd: Change in the Score of Similiar Documents
Hi, I have created two shards at SOLR Server and I have indexed 6 documents (all docs having exactly same data = Welcome to SOLR). Let’s say ids are from 1 to 6 and they are indexed in such a way : Shard_one : ids with 2,4,6 are present in this shard. Shard_two : ids with 1,3,5 are present in this shard. When I search “SOLR” , all documents are returned (as expected) but in order like 2, 4, 6, 1, 3, 5 With docs with id (2,4,6) having slightly high score than docs with id (1,3,5). I am not able to figure out why there is change in the score of docs from two different shards at the time of querying, even if the data in all docs are same. Is this because of indexing at multiple shards?? Please provide me some pointers to move ahead. Thanks, Rashi
Fwd: Reindex Issues
Hi, I have created two shards at solr server and around 4K documents are equally indexed over these two shards. I did re-indexing for all the indexed documents (updating exiting docs with same data again). After Re-indexing, I found that my indexes are not optimized and there is change in the score of documents. Previously when I Query *:* = returned all docs with same score (let’s say = 0.046) After Re-indexing, when I query *:* = returned all docs but with different and sudden high score (Let’s Say for first 3K docs = 0.46 and for last 1K = 0.45) I am not able to understand why there is sudden increase as well as change in the score of docs after re-indexing. Please provide me some pointers on this. Also, I wanted to know if I re-index small number of documents, why I am not able to see counts for deleted docs from UI and indexes are also optimized ?? Thanks, Rashi
Fwd: Issue with SOLR Distributed Search
Hi, This is regarding the issue that we are facing with SOLR distributed search. In our application, we are managing multiple shards at SOLR server to manage the load. But there is a problem with the order of results that we going to return to client during the search. For Example: Currently there are two shards on which data is randomly distributed. When I search something, it was observerd that the results from one shard appear first and then results from other shard. Moreover, we are ordering results by applying two levels of sorting (configurable as per user also): 1. Score 2. Modified Time I did investigations for the above scenario and found that it is not necessary that documents coming from one shard will always have the same score as documents coming from other shard, even if they are identical. I also went through the various SOLR documentations and links, and found that currently there is a limitation to distributed search in SOLR that Inverse-document frequency (IDF) calculations cannot be distributed and TF/IDF computations are per shard. This issue is particularly visible when there is significant difference between the number of documents indexed in each shard. (For Ex: first shard has 15000 docs and second shard has 5000). Please review and let me know whether our findings for the above scenario are appropriate or not. Also, as per our investigation currently there is work ongoing in SOLR community to support this concept of distributed/Global IDF. But, I wanted to know if there is any solution possible right now to manage/control the score of the documents during distributed search, so that the results seem more relevant. Thanks Rashi
Not able to deploy SOLR after applying OpenNLP patch
Hi, My Question is related to OpenNLP Integration with SOLR. I have successfully applied OpenNLP LUCENE-2899-x.patch to latest solr branch checkout from here: http://svn.apache.org/repos/asf/lucene/dev/branches/branch_4x And also iam able to compile source code, generated all realted binaries and able to create war file. But facing issues while deployment of SOLR. Here is the error Caused by: org.apache.solr.common.SolrException: Plugin init failure for [schema.xml] fieldType "text_opennlp": Plugin init failure for [schema.xml] a nalyzer/tokenizer: Error loading class 'solr.OpenNLPTokenizerFactory' at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177) at org.apache.solr.schema.IndexSchema.readSchema(IndexSchema.java:467) ... 15 more Caused by: org.apache.solr.common.SolrException: Plugin init failure for [schema.xml] analyzer/tokenizer: Error loading class 'solr.OpenNLPTokenizerFa ctory' at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177) at org.apache.solr.schema.FieldTypePluginLoader.readAnalyzer(FieldTypePluginLoader.java:362) at org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:95) at org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:43) at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:151) ... 16 more Caused by: org.apache.solr.common.SolrException: Error loading class 'solr.OpenNLPTokenizerFactory' at org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:449) at org.apache.solr.core.SolrResourceLoader.newInstance(SolrResourceLoader.java:543) at org.apache.solr.schema.FieldTypePluginLoader$2.create(FieldTypePluginLoader.java:342) at org.apache.solr.schema.FieldTypePluginLoader$2.create(FieldTypePluginLoader.java:335) at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:151) ... 20 more Caused by: java.lang.ClassNotFoundException: solr.OpenNLPTokenizerFactory at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:423) at java.net.FactoryURLClassLoader.loadClass(URLClassLoader.java:789) at java.lang.ClassLoader.loadClass(ClassLoader.java:356) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:264) at org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:433) ... 24 more 4446 [coreLoadExecutor-3-thread-1] ERROR org.apache.solr.core.CoreContainer û null:org.apache.solr.common.SolrException: Unable to create core: colle ction1 at org.apache.solr.core.CoreContainer.recordAndThrow(CoreContainer.java:931) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:563) at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:244) at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:236) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) Please help me on this. Waiting for your reply. Thanks in advance.
OpenNLP Analyzers not working properly
Hi, iam working on OpenNLP with SOLR. I have successfully applied the patch LUCENE-2899-x.patch to latest SOLR code branch_4x. I desgined some analyers based on OpenNLP filters and tokenziers and index some documnets on that fields. Searching on OpenNLP field is not constant. Not able to search on these OpenNLP designed fields in solr schema.xml properly. Also, how to use payloads for boosting the document. Please help me on this.
SOLR: Searching on OpenNLP fields is unstable
HI, I am working on OpenNLP integration with SOLR. I have successfully applied the patch (LUCENE-2899-x.patch) to latest SOLR source code (branch_4x). I have designed OpenNLP analyzer and index data to it. Analyzer declaration in schema.xml is as And field declared for this analyzer: Problem is here : When I search over this field Detail_Person, results are not constant. When I search Detail_Person:brett, it return one document But again when I fire the same query, it return zero document. Searching is not stable on OpenNLP field, sometimes it return documents and sometimes not but documents are there. And if I search on non OpenNLP fields, it is working properly, results are stable and correct. Please help me to make solr results consistent. Thanks in Advance.
Control Solr spellcheck functionality to provide suggestions for correct word
HI, I am working on Solr spellcheck feature, and I am using index based spellcheck dictionary as a source for spellcheck suggestions. I observed that collated results returned by spellcheck component, provide the suggestions for misspelled words, however also provide suggestions for correctly spelled word in query. For example, misspelled query - root priviladge to user *collated results (even suggestion includes the same) *- root privilege to user, room privilege to user, root privilege to users, rest privilege to user, root privilege to used It corrected word 'privilege' which was misspelled, however also provided suggestions for 'root' or 'user', which were already correct. is there a way , we can tell Solr not to provide suggestions for correct word, when using spellcheck feature. Please provide pointers.
handling stopwords for special scenarios
Hi All, We are using stopword filter factory at both index and search time, to omit the stopwords. However, for a one particular case, we are getting "here" as a search query and "here" is one the words in title/name representing our client. We are returning zero results as "here" is one of the English language stopwords which is getting omitted while indexing and searching both. One solution could be that I remove the "here" from list of stopwords, however does not look feasible. Is there any way where we can handle this kind of cases, where stopwrods are meant to be actual search term? Any leads would be appreciated.
Can we have multiple Spellcheck Components under /select handler
Hi All I am trying to configure multiple spellcheck components. I defined two searchComponents in my solrconfig.xml, let's say and . And added above both components under /select request handler with default required attributes. > elevator spellcheck < > str>spellcheck1 However, I can not see spellcheck1 coming in response. (when I set /select?spellcheck1=true&q=). Can't we configure multiple spellcheck components with different names in Solr.