Replication after re adding nodes to cluster (sleeping replicas)
I have a SolrCloud cluster holding 4 collections, each with with 3 shards and replication factor = 2. They all live on 2 machines, and I am currently using this setup for testing. However, i would like to connect this test setup to our live application, just for benchmarking and evaluating if it can handle the big qpm number. I am planning also to setup a new machine, and add new nodes manually, one more replica for each shard on the new machines, in case the first two have problems handling the big qpm. But what i would like to do is after I set up the new nodes, to shut down the new machine and only put it back in the cluster if it's needed. Thus, getting to the title of this mail: After re adding the 3rd machine to the cluster, will the replicas be automatically synced with the leader, or do i need to manually trigger this somehow ? Is there a better idea for having this sleeping replicas? I bet lots of people faced this problem, so a best practice must be out there. - Thanks, Michael -- View this message in context: http://lucene.472066.n3.nabble.com/Replication-after-re-adding-nodes-to-cluster-sleeping-replicas-tp4098764.html Sent from the Solr - User mailing list archive at Nabble.com.
Custom Plugin exception : Plugin init failure for [schema.xml]
Hi all, I have a problem with custom plugin development in solr 4.x versions. I have developed custom filter and trying to install it but I got following exception. Schema xml : Btw, I could not get any exception on 3.x versions. Any help will be appreciated. *Exception : * org.apache.solr.common.SolrException: Plugin init failure for [schema.xml] fieldType "text_tr": Plugin init failure for [schema.xml] analyzer/filter: Error instantiating class: 'my_package.TurkishFilterFactory' at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177) at org.apache.solr.schema.IndexSchema.readSchema(IndexSchema.java:374) at org.apache.solr.schema.IndexSchema.(IndexSchema.java:122) at org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:1018) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:1051) at org.apache.solr.core.CoreContainer$3.call(CoreContainer.java:634) at org.apache.solr.core.CoreContainer$3.call(CoreContainer.java:629) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:724) Caused by: org.apache.solr.common.SolrException: Plugin init failure for [schema.xml] analyzer/filter: Error instantiating class: 'my_package.TurkishFilterFactory' at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177) at org.apache.solr.schema.FieldTypePluginLoader.readAnalyzer(FieldTypePluginLoader.java:392) at org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:95) at org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:43) at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:151) ... 14 more Caused by: org.apache.solr.common.SolrException: Error instantiating class: 'my_package.TurkishFilterFactory' at org.apache.solr.core.SolrResourceLoader.newInstance(SolrResourceLoader.java:473) at org.apache.solr.util.plugin.AbstractPluginLoader.create(AbstractPluginLoader.java:89) at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:151) ... 18 more Caused by: java.lang.InstantiationException: my_package.TurkishFilterFactory at java.lang.Class.newInstance(Class.java:359) at org.apache.solr.core.SolrResourceLoader.newInstance(SolrResourceLoader.java:470) ... 20 more Nov 01, 2013 11:59:27 AM org.apache.solr.common.SolrException log SEVERE: null:org.apache.solr.common.SolrException: Unable to create core: collection1 at org.apache.solr.core.CoreContainer.recordAndThrow(CoreContainer.java:1672) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:1057) at org.apache.solr.core.CoreContainer$3.call(CoreContainer.java:634) at org.apache.solr.core.CoreContainer$3.call(CoreContainer.java:629) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:724) Caused by: org.apache.solr.common.SolrException: Plugin init failure for [schema.xml] fieldType "text_tr": Plugin init failure for [schema.xml] analyzer/filter: Error instantiating class: 'my_package.TurkishFilterFactory' at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177) at org.apache.solr.schema.IndexSchema.readSchema(IndexSchema.java:374) at org.apache.solr.schema.IndexSchema.(IndexSchema.java:122) at org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:1018) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:1051) ... 10 more Caused by: org.apache.solr.common.SolrException: Plugin init failure for [schema.xml] analyzer/filter: Error instantiating class: 'my_package.TurkishFilterFactory' at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177) at org.apache.solr.schema.FieldTypePluginLoader.readAnalyzer(FieldTypePluginLoader.java:392) at org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:95) at org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:43) at org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.j
Re: Store Solr OpenBitSets In Solr Indexes
Why are you saving this? Because if the bitset you're saving has anything to do with, say, filter queries, it's probably useless. The internal bitsets are often based on the internal Lucene doc ID, which will change when segment merges happen, thus the caution. Otherwise, theres the binary type you can probably use. It's not very efficient since I believe it uses base-64 encoding under the covers though... Is this an "XY" problem? Best, Erick On Wed, Oct 30, 2013 at 8:06 AM, David Philip wrote: > Hi All, > > What should be the field type if I have to save solr's open bit set value > within solr document object and retrieve it later for search? > > OpenBitSet bits = new OpenBitSet(); > > bits.set(0); > bits.set(1000); > > doc.addField("SolrBitSets", bits); > > > What should be the field type of SolrBitSets? > > Thanks >
Re: Indexing logs files of thousands of GBs
Throwing a multi-gigabyte file at Solr and expecting it to index it is asking for a bit too much. You either have to stream it up and break it apart or something similar. And consider what happens if you just index the log as a single document. How do you search it? Do you return several G as the result? Most applications break the log file up into individual documents and index each event individually to enable searches like "all OOM errors between 12:00 and 13:00 yesterday" or similar. How do you expect to do such a thing if it's one big document? I may be completely off base here, but I think you need to define the problem you're solving more clearly. I can flat guarantee that trying to index a large log file as one document will be unsatisfactory to search, even if you can get it into the index. Best, Erick On Wed, Oct 30, 2013 at 12:47 PM, Otis Gospodnetic < otis.gospodne...@gmail.com> wrote: > Hi, > > Hm, sorry for not helping with this particular issue directly, but it > looks like you are *uploading* your logs and indexing that way? > Wouldn't pushing them be a better fit when it comes to log indexing? > We recently contributed a Logstash output that can index logs to Solr, > which may be of interest - have a look at > https://twitter.com/otisg/status/395563043045638144 -- includes a > little diagram that shows how this fits into the picture. > > Otis > -- > Performance Monitoring * Log Analytics * Search Analytics > Solr & Elasticsearch Support * http://sematext.com/ > > > > On Wed, Oct 30, 2013 at 9:55 AM, keshari.prerna > wrote: > > Hello, > > > > As suggested by Chris, now I am accessing the files using java program > and > > creating SolrInputDocument, but i ran into this exception while doing > > server.add(document). When i tried to increase "ramBufferSizeMB", it > doesn't > > let me make it more than 2 gig. > > > > org.apache.solr.client.solrj.SolrServerException: Server at > > http://localhost:8983/solr/logsIndexing returned non ok status:500, > > message:the request was rejected because its size (2097454) exceeds the > > configured maximum (2097152) > > org.apache.commons.fileupload.FileUploadBase$SizeLimitExceededException: > the > > request was rejected because its size (2097454) exceeds the configured > > maximum (2097152) at > > > org.apache.commons.fileupload.FileUploadBase$FileItemIteratorImpl$1.raiseError(FileUploadBase.java:902) > > at > > > org.apache.commons.fileupload.util.LimitedInputStream.checkLimit(LimitedInputStream.java:71) > > at > > > org.apache.commons.fileupload.util.LimitedInputStream.read(LimitedInputStream.java:128) > > at > > > org.apache.commons.fileupload.MultipartStream$ItemInputStream.makeAvailable(MultipartStream.java:977) > > at > > > org.apache.commons.fileupload.MultipartStream$ItemInputStream.read(MultipartStream.java:887) > > at java.io.InputStream.read(Unknown Source) at > > org.apache.commons.fileupload.util.Streams.copy(Streams.java:94) > at > > org.apache.commons.fileupload.util.Streams.copy(Streams.java:64) > at > > > org.apache.commons.fileupload.FileUploadBase.parseRequest(FileUploadBase.java:362) > > at > > > org.apache.commons.fileupload.servlet.ServletFileUpload.parseRequest(ServletFileUpload.java:126) > > at > > > org.apache.solr.servlet.MultipartRequestParser.parseParamsAndFillStreams(SolrRequestParsers.java:344) > > at > > > org.apache.solr.servlet.StandardRequestParser.parseParamsAndFillStreams(SolrRequestParsers.java:397) > > at > > > org.apache.solr.servlet.SolrRequestParsers.parse(SolrRequestParsers.java:115) > > at > > > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:244) > > at > > > org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212) > > at > org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399) > > at > > > org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) > > at > org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182) > > at > org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766) > > at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450) > > at org.mortbay.jetty.handler.ContextHand > > at > > > org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:328) > > at > > > org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:211) > > at > > > org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105) > > at > org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:121) > > at > org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:106) > > at Filewalker.walk(LogsIndexer.java:48) > > at Filewalker.main(LogsIndexer.java:69) > > > > How do I get rid of this? > > > > Thanks, > > Prerna > > > > > > > > -- > > View this message in context: > http://lucene.472066.n3.nabble.com/Indexing-logs-files-of-thousands-of-GBs-
Re: Simple (?) zookeeper question
Unknown document router errors are usually caused by using different solr and solrj versions - which version of solr and solrj are you using? Alan Woodward www.flax.co.uk On 1 Nov 2013, at 04:19, Jack Park wrote: > After digging deeper (slow for a *nix newbee), I uncovered issues with > the java installation. A step in installation of Oracle Java has it > that you -install "java" with the path to /bin/java. That done, > zookeeper seems to be running. > > I booted three cores (on the same box) -- this is the simple one-box > 3-node cloud test, and used the test code from the Lucidworks course > to send over and read some documents. That failed with this: > Unknown document router '{name=compositeId}' > > Lots more research. > Closer... > > On Thu, Oct 31, 2013 at 5:44 PM, Jack Park wrote: >> Latest zookeeper is installed on an Ubuntu server box. >> Java is 1.7 latest build. >> whereis points to java just fine. >> /etc/zookeeper is empty. >> >> boot zookeeper from /bin as sudo ./zkServer.sh start >> Console says "Started" >> /etc/zookeeper now has a .pid file >> In another console, ./zkServer.sh status returns: >> "It's probably not running" >> >> An interesting fact: the log4j.properties file says there should be a >> zookeeper.log file in "."; there is no log file. When I do a text >> search in the zookeeper source code for where it picks up the >> log4j.properties, nothing is found. >> >> Fascinating, what? This must be a common beginner's question, not >> well covered in web-search for my context. Does it ring any bells? >> >> Many thanks. >> Jack
Re: How to get similarity score between 0 and 1 not relative score
There's another use case for scaling the score. Suppose I want to compute a custom score based on the weighted sum of: - product(0.75, relevance score) - product(0.25, value from another field) For this to work, both fields must have values between 0-1, for example. Toby's example using the scale function seems to work, but you have to use fq to eliminate results with score=0. It seems this is somewhat expensive, since the scaling can't be done until all results have been collected to get the max score. Then, are the results resorted? I haven't looked closely, yet. Peter Peter On Thu, Oct 31, 2013 at 7:48 PM, Toby Lazar wrote: > I think you are looking for something like this, though you can omit the fq > section: > > > > http://localhost:8983/solr/collection/select?abc=text:bob&q={!func}scale(product(query($abc),1),0,1)&fq={ > ! > frange l=0.9}$q > > Also, I don't understand all the fuss about normalized scores. In the > linked example, I can see an interest in searching for "apple bannana", > "zzz yyy xxx qqq kkk ttt rrr 111", etc. and wanting only close matches for > that point in time. Would this be a good use for this approach? I > understand that the results can change if the documents in the index > change. > > Thanks, > > Toby > > > > On Thu, Oct 31, 2013 at 12:56 AM, Anshum Gupta >wrote: > > > Hi Susheel, > > > > Have a look at this: > > http://wiki.apache.org/lucene-java/ScoresAsPercentages > > > > You may really want to reconsider doing that. > > > > > > > > > > On Thu, Oct 31, 2013 at 9:41 AM, sushil sharma > >wrote: > > > > > Hi, > > > > > > We have a requirement where user would like to see a score (between 0 > to > > > 1) which can tell how close the input search string is with result > > string. > > > So if input was very close but not exact matach, score could be .90 > etc. > > > > > > I do understand that we can get score from solr & divide by highest > score > > > but that will always show 1 even if we match was not exact. > > > > > > Regards, > > > Susheel > > > > > > > > > > -- > > > > Anshum Gupta > > http://www.anshumgupta.net > > >
Can I combine standardtokenizer with solr.WordDelimiterFilterFactory?
Good morning, Here's the issue: I have and ID that consists of two letters and a number. The whole user title looks like this: Lastname, Firstname (LA12345). Now, with my current configuration, I can search for LA12345 and find the user. However, when I type in just the number I get zero results. If I put a wildcard in (*12345) I find the correct record. The problem is I changed that user title to use the worddelimiterfitlerfactory and it seems to work. However, I also copy that field into the text field which just uses the standardtokenizer and I lose the ability to search for 12345 without a wildcard. My question is can (or should) I put the worddelimiterfactory in with the standardtokenizer in the text field? Or should I just use one or the other? Thank you, -- View this message in context: http://lucene.472066.n3.nabble.com/Can-I-combine-standardtokenizer-with-solr-WordDelimiterFilterFactory-tp4098814.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: HTTP Basic Authentication with solr's jetty
Thus wrote Fabiano Sidler: > Hi folks! > > I was asking this question last week already on the jetty mailing list, but > haven't got any answer. > > I'm going to run multiple Solr instances on one server, which arises the need > of user authentication in front of Solr. I've done the following steps (after > a lot of others which didn't work): > > === snip === > diff -wur solr-4.5.0.orig/example/etc/jetty.xml > solr-4.5.0/example/etc/jetty.xml > --- solr-4.5.0.orig/example/etc/jetty.xml 2013-09-12 14:12:53.0 +0200 > +++ solr-4.5.0/example/etc/jetty.xml 2013-10-28 17:14:49.0 +0100 > @@ -12,6 +12,16 @@ > > > > + > + > + > + Test Realm > + name="config">/some/path/solr-4.5.0/example/passwords.properties > + 0 > + > + > + > + > > > > diff -wur solr-4.5.0.orig/example/etc/webdefault.xml > solr-4.5.0/example/etc/webdefault.xml > --- solr-4.5.0.orig/example/etc/webdefault.xml2013-09-12 > 14:12:53.0 +0200 > +++ solr-4.5.0/example/etc/webdefault.xml2013-10-28 > 17:27:43.0 +0100 > @@ -520,8 +520,14 @@ >/ >TRACE > > - > + > + * > + > > > + > +BASIC > +Test Realm > + > > diff > === snap === > > The example/password.properties file contains the following test accounts: > === snip === > user: user,user > admin: admin,admin > === snap === > > The problem now is that still no authentication is run by jetty and thus I > can connect to Solr core without being authenticated. What is still neede > to finally activate authentication? > > Greetings, > Fabiano Hello? Not anyone with an idea or who has already done that? Greetings, Fabiano
Re: Can I combine standardtokenizer with solr.WordDelimiterFilterFactory?
Yes you can and I don't see any reason why you should not Le 01/11/2013 15:38, eShard a écrit : Good morning, Here's the issue: I have and ID that consists of two letters and a number. The whole user title looks like this: Lastname, Firstname (LA12345). Now, with my current configuration, I can search for LA12345 and find the user. However, when I type in just the number I get zero results. If I put a wildcard in (*12345) I find the correct record. The problem is I changed that user title to use the worddelimiterfitlerfactory and it seems to work. However, I also copy that field into the text field which just uses the standardtokenizer and I lose the ability to search for 12345 without a wildcard. My question is can (or should) I put the worddelimiterfactory in with the standardtokenizer in the text field? Or should I just use one or the other? Thank you, -- View this message in context: http://lucene.472066.n3.nabble.com/Can-I-combine-standardtokenizer-with-solr-WordDelimiterFilterFactory-tp4098814.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Can I combine standardtokenizer with solr.WordDelimiterFilterFactory?
On Fri, Nov 1, 2013 at 11:38 AM, eShard wrote: > I have and ID that consists of two letters and a number. > The whole user title looks like this: Lastname, Firstname (LA12345). > Now, with my current configuration, I can search for LA12345 and find the > user. > However, when I type in just the number I get zero results. > If I put a wildcard in (*12345) I find the correct record. > The problem is I changed that user title to use the > worddelimiterfitlerfactory and it seems to work. > However, I also copy that field into the text field which just uses the > standardtokenizer and I lose the ability to search for 12345 without a > wildcard. > My question is can (or should) I put the worddelimiterfactory in with the > standardtokenizer in the text field? copyField worlks before any analysis, so if you want the splitting behavior on your "text" field, then you'll need to use the WordDelimiterFilter there also. -Yonik http://heliosearch.com -- making solr shine
Re: Simple (?) zookeeper question
Alan, That was brilliant! My test harness was behind a couple of notches. Hah! So, now we open yet another can of strange looking creatures, namely: No live SolrServers available to handle this request:[http://127.0.1.1:8983/solr/collection1] at org.apache.solr.client.solrj.impl.CloudSolrServer.directUpdate(CloudSolrServer.java:347) 3 times, once for each URL I passed into the server. Here is the code: String zkurl = "10.1.10.178:2181"; String solrurla = "10.1.10.178:8983"; String solrurlb = "10.1.10.178:7574"; String solrurlc = "10.1.10.178:7590"; LBHttpSolrServer sv = new LBHttpSolrServer(solrurla,solrurlb,solrurlc); CloudSolrServer server = new CloudSolrServer(zkurl,sv); server.setDefaultCollection("collection1"); I am struggling to imagine how 10.1.10.178 got translated to 127.0.1.1 and the port assignments ignored for each URL passed in. That error message seems well known to search engines. One suggestion is to check the zookeeper logs. According to the zookeeper's log4j properties, there should be a zookeeper.log in the zookeeper directory. There is no such log. I went to /etc/zookeeper/Version_2 and looked at log.1 (binary) but could see hints that this might be where the 127.0.1.1 is coming from: zookeeper sending such an error message back. This would suggest that, somehow or other, my nodes are not properly registering themselves, though no error messages were tossed when each node was booted. solr.log for node1 only reflects queries from the admin page. That's what I am working on now. Thanks! On Fri, Nov 1, 2013 at 6:03 AM, Alan Woodward wrote: > Unknown document router errors are usually caused by using different solr and > solrj versions - which version of solr and solrj are you using? > > Alan Woodward > www.flax.co.uk > > > On 1 Nov 2013, at 04:19, Jack Park wrote: > >> After digging deeper (slow for a *nix newbee), I uncovered issues with >> the java installation. A step in installation of Oracle Java has it >> that you -install "java" with the path to /bin/java. That done, >> zookeeper seems to be running. >> >> I booted three cores (on the same box) -- this is the simple one-box >> 3-node cloud test, and used the test code from the Lucidworks course >> to send over and read some documents. That failed with this: >> Unknown document router '{name=compositeId}' >> >> Lots more research. >> Closer... >> >> On Thu, Oct 31, 2013 at 5:44 PM, Jack Park wrote: >>> Latest zookeeper is installed on an Ubuntu server box. >>> Java is 1.7 latest build. >>> whereis points to java just fine. >>> /etc/zookeeper is empty. >>> >>> boot zookeeper from /bin as sudo ./zkServer.sh start >>> Console says "Started" >>> /etc/zookeeper now has a .pid file >>> In another console, ./zkServer.sh status returns: >>> "It's probably not running" >>> >>> An interesting fact: the log4j.properties file says there should be a >>> zookeeper.log file in "."; there is no log file. When I do a text >>> search in the zookeeper source code for where it picks up the >>> log4j.properties, nothing is found. >>> >>> Fascinating, what? This must be a common beginner's question, not >>> well covered in web-search for my context. Does it ring any bells? >>> >>> Many thanks. >>> Jack >
Re: Custom Plugin exception : Plugin init failure for [schema.xml]
On 11/1/2013 4:18 AM, Parvin Gasimzade wrote: > I have a problem with custom plugin development in solr 4.x versions. I > have developed custom filter and trying to install it but I got following > exception. Later you indicated that you can use it with Solr 3.x without any problem. Did you recompile your custom plugin against the Solr jars from the new version? There was a *huge* amount of java class refactoring that went into the 4.0 version as compared to any 3.x version, and that continues with each new 4.x release. I would bet that if you tried that recompile, it would fail due to errors and/or warnings, which you'll need to fix. There might also be operational problems that the compiler doesn't find, due to changes in how the underlying APIs get used. Thanks, Shawn
Exclude urls without 'www' from Nutch 1.7 crawl
I'm currently using Nutch 1.7 to crawl my domain. My issue is specific to URLs being indexed as www vs. non-www. Specifically, after firing the crawl and index to Solr 4.5 then validating the results on the front-end with AJAX Solr, the search results page lists results/pages that are both 'www' and '' urls such as: www.mywebsite.com mywebsite.com www.mywebsite.com/page1 mywebsite.com/page1 My understanding is that the url filtering (regex-urlfilter.txt) needs modification. Are there any regex/nutch experts that could suggest a solution? Here is the code on paste bin, http://pastebin.com/Cp6vUxPR Also on stack overflow, http://stackoverflow.com/questions/19731904/exclude-urls-without-www-from-nutch-1-7-crawl Thank you, Mark IMPORTANT NOTICE: This e-mail message is intended to be received only by persons entitled to receive the confidential information it may contain. E-mail messages sent from Bridgepoint Education may contain information that is confidential and may be legally privileged. Please do not read, copy, forward or store this message unless you are an intended recipient of it. If you received this transmission in error, please notify the sender by reply e-mail and delete the message and any attachments.
Re: Simple (?) zookeeper question
/clusterstate.json seems to clearly state that all 3 nodes are alive, have ranges, and are active. Still, it would seem that java is still not properly installed. ZooKeeper is dropping zookeeper.out in the /bin directory, which says this, among other things: Server environment:java.home=/usr/local/java/jdk1.7.0_40/jre Server environment:java.class.path=/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../build/classes:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../build/lib/*.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/slf4j-log4j12-1.6.1.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/slf4j-api-1.6.1.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/netty-3.2.2.Final.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/log4j-1.2.15.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/jline-0.9.94.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../zookeeper-3.4.5.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../src/java/lib/*.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../conf: Server environment:java.library.path= /usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib There is no /usr/java/... It's really a mystery where zookeeper is getting these values; everything else seems right. But, for me, here's the amazing chunk of traces (cleaned up a bit) Accepted socket connection from /127.0.0.1:39065 Client attempting to establish new session at /127.0.0.1:39065 Established session 0x1421197e6e90002 with negotiated timeout 15000 for client /127.0.0.1:39065 Got user-level KeeperException when processing sessionid:0x1421197e6e90002 type:create cxid:0x1 zxid:0xc0 txntype:-1 reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists for /overseer Got user-level KeeperException when processing sessionid:0x1421197e6e90002 type:create cxid:0x3 zxid:0xc1 txntype:-1 reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists for /overseer Got user-level KeeperException when processing sessionid:0x1421197e6e90002 type:delete cxid:0xe zxid:0xc2 txntype:-1 reqpath:n/a Error Path:/live_nodes/127.0.1.1:7590_solr Error:KeeperErrorCode = NoNode for /live_nodes/127.0.1.1:7590_solr Got user-level KeeperException when processing sessionid:0x1421197e6e90002 type:delete cxid:0x9f zxid:0xcd txntype:-1 reqpath:n/a Error Path:/collections/collection1/leaders/shard3 Error:KeeperErrorCode = NoNode for /collections/collection1/leaders/shard3 2013-10-31 21:01:19,344 [myid:] - INFO [ProcessThread(sid:0 cport:-1)::PrepRequestProcessor@627] - Got user-level KeeperException when processing sessionid:0x1421197e6e90002 type:create cxid:0xa0 zxid:0xce txntype:-1 reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists for /overseer Got user-level KeeperException when processing sessionid:0x1421197e6e90002 type:create cxid:0xaa zxid:0xd1 txntype:-1 reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists for /overseer Accepted socket connection from /10.1.10.180:55528 Client attempting to establish new session at /10.1.10.180:55528 Established session 0x1421197e6e90003 with negotiated timeout 1 for client /10.1.10.180:55528 WARN Exception causing close of session 0x1421197e6e90003 due to java.io.IOException: Connection reset by peer Closed socket connection for client /10.1.10.180:55528 which had sessionid 0x1421197e6e90003 Sockets from 10.1.10.180 are my windoz box shipping solr documents. I am not sure how I am using 55528 unless that's a solrj behavior. Connection reset by peer would suggest something in my code, but my code is a clone of code supplied in a Solr training course. Must be good. Right? I also have no clue what is /127.0.0.1:39065 -- that's not one of my nodes. The quest continues. On Fri, Nov 1, 2013 at 9:21 AM, Jack Park wrote: > Alan, > That was brilliant! > My test harness was behind a couple of notches. > > Hah! So, now we open yet another can of strange looking creatures, namely: > > No live SolrServers available to handle this > request:[http://127.0.1.1:8983/solr/collection1] > at > org.apache.solr.client.solrj.impl.CloudSolrServer.directUpdate(CloudSolrServer.java:347) > > 3 times, once for each URL I passed into the server. Here is the code: > > String zkurl = "10.1.10.178:2181"; > String solrurla = "10.1.10.178:8983"; > String solrurlb = "10.1.10.178:7574"; > String solrurlc = "10.1.10.178:7590"; > > LBHttpSolrServer sv = new LBHttpSolrServer(solrurla,solrurlb,solrurlc); > CloudSolrServer server = new CloudSolrServer(zkurl,sv); > server.setDefaultCollection("collection1"); > > I am struggling to imagine how 10.1.10.178 got translated to 127.0.1.1 > and the port assignments ignored for each URL passed in. > > That error message seems well known to search engines. One suggestion > is to check the zookeeper logs. According to the zookeeper's log4j > properties, there should be a zookeeper.log in the zookeeper > directory.
RE: Exclude urls without 'www' from Nutch 1.7 crawl
Hi - Use the domain-urlfilter for host, domain and TLD filtering. Also, please ask questions on the Nutch list, you're on Solr now :) -Original message- > From:Reyes, Mark > Sent: Friday 1st November 2013 17:24 > To: solr-user@lucene.apache.org > Subject: Exclude urls without 'www' from Nutch 1.7 crawl > > I'm currently using Nutch 1.7 to crawl my domain. My issue is specific to > URLs being indexed as www vs. non-www. > > Specifically, after firing the crawl and index to Solr 4.5 then validating > the results on the front-end with AJAX Solr, the search results page lists > results/pages that are both 'www' and '' urls such as: > > www.mywebsite.com > mywebsite.com > www.mywebsite.com/page1 > mywebsite.com/page1 > > My understanding is that the url filtering (regex-urlfilter.txt) needs > modification. Are there any regex/nutch experts that could suggest a solution? > > Here is the code on paste bin, > http://pastebin.com/Cp6vUxPR > > Also on stack overflow, > http://stackoverflow.com/questions/19731904/exclude-urls-without-www-from-nutch-1-7-crawl > > Thank you, > Mark > > > IMPORTANT NOTICE: This e-mail message is intended to be received only by > persons entitled to receive the confidential information it may contain. > E-mail messages sent from Bridgepoint Education may contain information that > is confidential and may be legally privileged. Please do not read, copy, > forward or store this message unless you are an intended recipient of it. If > you received this transmission in error, please notify the sender by reply > e-mail and delete the message and any attachments.
Re: Exclude urls without 'www' from Nutch 1.7 crawl
As Markus pointed Nutch has a feature for such kind of situation. Here is Solr list but one more thing for you: www.mywebsite.com and mywebsite.commay point to "different" pages. 2013/11/1 Markus Jelsma > Hi - Use the domain-urlfilter for host, domain and TLD filtering. > > Also, please ask questions on the Nutch list, you're on Solr now :) > > > -Original message- > > From:Reyes, Mark > > Sent: Friday 1st November 2013 17:24 > > To: solr-user@lucene.apache.org > > Subject: Exclude urls without 'www' from Nutch 1.7 crawl > > > > I'm currently using Nutch 1.7 to crawl my domain. My issue is specific > to URLs being indexed as www vs. non-www. > > > > Specifically, after firing the crawl and index to Solr 4.5 then > validating the results on the front-end with AJAX Solr, the search results > page lists results/pages that are both 'www' and '' urls such as: > > > > www.mywebsite.com > > mywebsite.com > > www.mywebsite.com/page1 > > mywebsite.com/page1 > > > > My understanding is that the url filtering (regex-urlfilter.txt) needs > modification. Are there any regex/nutch experts that could suggest a > solution? > > > > Here is the code on paste bin, > > http://pastebin.com/Cp6vUxPR > > > > Also on stack overflow, > > > http://stackoverflow.com/questions/19731904/exclude-urls-without-www-from-nutch-1-7-crawl > > > > Thank you, > > Mark > > > > > > IMPORTANT NOTICE: This e-mail message is intended to be received only by > persons entitled to receive the confidential information it may contain. > E-mail messages sent from Bridgepoint Education may contain information > that is confidential and may be legally privileged. Please do not read, > copy, forward or store this message unless you are an intended recipient of > it. If you received this transmission in error, please notify the sender by > reply e-mail and delete the message and any attachments. >
Re: Exclude urls without 'www' from Nutch 1.7 crawl
Noted and will do (that goes twice for the suggestions and putting this on the nutch list instead). Thanks all, Mark On 11/1/13, 10:53 AM, "Furkan KAMACI" wrote: >As Markus pointed Nutch has a feature for such kind of situation. Here is >Solr list but one more thing for you: www.mywebsite.com and >mywebsite.commay point to "different" pages. > > >2013/11/1 Markus Jelsma > >> Hi - Use the domain-urlfilter for host, domain and TLD filtering. >> >> Also, please ask questions on the Nutch list, you're on Solr now :) >> >> >> -Original message- >> > From:Reyes, Mark >> > Sent: Friday 1st November 2013 17:24 >> > To: solr-user@lucene.apache.org >> > Subject: Exclude urls without 'www' from Nutch 1.7 crawl >> > >> > I'm currently using Nutch 1.7 to crawl my domain. My issue is specific >> to URLs being indexed as www vs. non-www. >> > >> > Specifically, after firing the crawl and index to Solr 4.5 then >> validating the results on the front-end with AJAX Solr, the search >>results >> page lists results/pages that are both 'www' and '' urls such as: >> > >> > www.mywebsite.com >> > mywebsite.com >> > www.mywebsite.com/page1 >> > mywebsite.com/page1 >> > >> > My understanding is that the url filtering (regex-urlfilter.txt) needs >> modification. Are there any regex/nutch experts that could suggest a >> solution? >> > >> > Here is the code on paste bin, >> > http://pastebin.com/Cp6vUxPR >> > >> > Also on stack overflow, >> > >> >>http://stackoverflow.com/questions/19731904/exclude-urls-without-www-from >>-nutch-1-7-crawl >> > >> > Thank you, >> > Mark >> > >> > >> > IMPORTANT NOTICE: This e-mail message is intended to be received only >>by >> persons entitled to receive the confidential information it may contain. >> E-mail messages sent from Bridgepoint Education may contain information >> that is confidential and may be legally privileged. Please do not read, >> copy, forward or store this message unless you are an intended >>recipient of >> it. If you received this transmission in error, please notify the >>sender by >> reply e-mail and delete the message and any attachments. >> IMPORTANT NOTICE: This e-mail message is intended to be received only by persons entitled to receive the confidential information it may contain. E-mail messages sent from Bridgepoint Education may contain information that is confidential and may be legally privileged. Please do not read, copy, forward or store this message unless you are an intended recipient of it. If you received this transmission in error, please notify the sender by reply e-mail and delete the message and any attachments.
Re: Simple (?) zookeeper question
The top error message at my test harness is this: No live SolrServers available to handle this request: [http://127.0.1.1:8983/solr/collection1, http://127.0.1.1:7574/solr/collection1, http://127.0.1.1:7590/solr/collection1] I have to assume that error message was somehow shipped by zookeeper, because those servers actually exist, to the test harness, at 10.1.10.178, and if I access any one of them from the browser, /solr/collection1 does not work, but /solr/#/collection1 does work. On Fri, Nov 1, 2013 at 10:34 AM, Jack Park wrote: > /clusterstate.json seems to clearly state that all 3 nodes are alive, > have ranges, and are active. > > Still, it would seem that java is still not properly installed. > ZooKeeper is dropping zookeeper.out in the /bin directory, which says > this, among other things: > > Server environment:java.home=/usr/local/java/jdk1.7.0_40/jre > > Server > environment:java.class.path=/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../build/classes:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../build/lib/*.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/slf4j-log4j12-1.6.1.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/slf4j-api-1.6.1.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/netty-3.2.2.Final.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/log4j-1.2.15.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/jline-0.9.94.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../zookeeper-3.4.5.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../src/java/lib/*.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../conf: > > Server environment:java.library.path= > /usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib > > There is no /usr/java/... > It's really a mystery where zookeeper is getting these values; > everything else seems right. > > But, for me, here's the amazing chunk of traces (cleaned up a bit) > > Accepted socket connection from /127.0.0.1:39065 > Client attempting to establish new session at /127.0.0.1:39065 > Established session 0x1421197e6e90002 with negotiated timeout 15000 > for client /127.0.0.1:39065 > Got user-level KeeperException when processing > sessionid:0x1421197e6e90002 type:create cxid:0x1 zxid:0xc0 txntype:-1 > reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists > for /overseer > Got user-level KeeperException when processing > sessionid:0x1421197e6e90002 type:create cxid:0x3 zxid:0xc1 txntype:-1 > reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists > for /overseer > Got user-level KeeperException when processing > sessionid:0x1421197e6e90002 type:delete cxid:0xe zxid:0xc2 txntype:-1 > reqpath:n/a Error Path:/live_nodes/127.0.1.1:7590_solr > Error:KeeperErrorCode = NoNode for /live_nodes/127.0.1.1:7590_solr > Got user-level KeeperException when processing > sessionid:0x1421197e6e90002 type:delete cxid:0x9f zxid:0xcd txntype:-1 > reqpath:n/a Error Path:/collections/collection1/leaders/shard3 > Error:KeeperErrorCode = NoNode for > /collections/collection1/leaders/shard3 > 2013-10-31 21:01:19,344 [myid:] - INFO [ProcessThread(sid:0 > cport:-1)::PrepRequestProcessor@627] - Got user-level KeeperException > when processing sessionid:0x1421197e6e90002 type:create cxid:0xa0 > zxid:0xce txntype:-1 reqpath:n/a Error Path:/overseer > Error:KeeperErrorCode = NodeExists for /overseer > Got user-level KeeperException when processing > sessionid:0x1421197e6e90002 type:create cxid:0xaa zxid:0xd1 txntype:-1 > reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists > for /overseer > Accepted socket connection from /10.1.10.180:55528 > Client attempting to establish new session at /10.1.10.180:55528 > Established session 0x1421197e6e90003 with negotiated timeout 1 > for client /10.1.10.180:55528 > WARN Exception causing close of session 0x1421197e6e90003 due to > java.io.IOException: Connection reset by peer > Closed socket connection for client /10.1.10.180:55528 which had > sessionid 0x1421197e6e90003 > > Sockets from 10.1.10.180 are my windoz box shipping solr documents. I > am not sure how I am using 55528 unless that's a solrj behavior. > Connection reset by peer would suggest something in my code, but my > code is a clone of code supplied in a Solr training course. Must be > good. Right? > > I also have no clue what is /127.0.0.1:39065 -- that's not one of my nodes. > > The quest continues. > > On Fri, Nov 1, 2013 at 9:21 AM, Jack Park wrote: >> Alan, >> That was brilliant! >> My test harness was behind a couple of notches. >> >> Hah! So, now we open yet another can of strange looking creatures, namely: >> >> No live SolrServers available to handle this >> request:[http://127.0.1.1:8983/solr/collection1] >> at >> org.apache.solr.client.solrj.impl.CloudSolrServer.directUpdate(CloudSolrServer.java:347) >> >> 3 times, once for each URL I passed into the server. Here is the code: >> >> S
Re: Simple (?) zookeeper question
On 11/1/2013 12:07 PM, Jack Park wrote: The top error message at my test harness is this: No live SolrServers available to handle this request: [http://127.0.1.1:8983/solr/collection1, http://127.0.1.1:7574/solr/collection1, http://127.0.1.1:7590/solr/collection1] I have to assume that error message was somehow shipped by zookeeper, because those servers actually exist, to the test harness, at 10.1.10.178, and if I access any one of them from the browser, /solr/collection1 does not work, but /solr/#/collection1 does work. Those are *base* urls. By themselves, they return 404. For an example of how a base URL is used, try /solr/collection1/select?q=*:* instead. Any URL with /#/ in it is part of the admin UI, which runs mostly in the browser and accesses Solr handlers to gather information. It is not Solr itself. Thanks, Shawn
Re: Simple (?) zookeeper question
Thanks. I reviewed clusterstate.json again; those URLs are alive. Why they are not responding seems to be the mystery du jour. I reviewed my test suite: it is using field names in schema.xml, and the server is configured to use the update responders I installed, all of which work fine in a non-cloud mode. Thanks Jack On Fri, Nov 1, 2013 at 11:12 AM, Shawn Heisey wrote: > On 11/1/2013 12:07 PM, Jack Park wrote: >> >> The top error message at my test harness is this: >> >> No live SolrServers available to handle this request: >> [http://127.0.1.1:8983/solr/collection1, >> http://127.0.1.1:7574/solr/collection1, >> http://127.0.1.1:7590/solr/collection1] >> >> I have to assume that error message was somehow shipped by zookeeper, >> because those servers actually exist, to the test harness, at >> 10.1.10.178, and if I access any one of them from the browser, >> /solr/collection1 does not work, but /solr/#/collection1 does work. > > > Those are *base* urls. By themselves, they return 404. For an example of > how a base URL is used, try /solr/collection1/select?q=*:* instead. > > Any URL with /#/ in it is part of the admin UI, which runs mostly in the > browser and accesses Solr handlers to gather information. It is not Solr > itself. > > Thanks, > Shawn >
Re: Error with CollapsingQParserPlugin when trying to use tagging
The following unit test in TestCollapseQParserPlugin.java illustrates the bug ModifiableSolrParams params = new ModifiableSolrParams(); params.add("q", "*:*"); params.add("fq", "{!collapse field=group_s}"); params.add("defType", "edismax"); params.add("bf", "field(test_ti)"); params.add("fq","{!tag=test_ti}test_ti:5"); params.add("facet","true"); params.add("facet.field","{!ex=test_ti}test_ti"); assertQ(req(params), "*[count(//doc)=1]", "//doc[./int[@name='test_ti']='5']") On Thu, Oct 31, 2013 at 6:46 PM, dboychuck [via Lucene] < ml-node+s472066n4098710...@n3.nabble.com> wrote: > Here is an example URL that gives the error: > > solr/productindex/productQuery?fq={!collapse%20field=groupid}&fq=discontinued:false&fq={!tag=manufacturer_string}manufacturer_string:(%22delta%22%20OR%20%22kohler%22)&fq=siteid:82&sort=score%20desc&facet=true&start=0&rows=48&fl=productid,manufacturer,uniqueFinish,uniqueid,productCompositeid,score&facet.query={!ex=onSale}onSale:true&facet.query={!ex=rating}rating:[4%20TO%20*]&facet.query={!ex=rating}rating:[3%20TO%20*]&facet.query={!ex=rating}rating:[2%20TO%20*]&facet.query={!ex=rating}rating:[1%20TO%20*]&facet.query={!ex=MadeinAmerica_boolean}MadeinAmerica_boolean:yes&facet.query={!ex=inStock}inStock:true&facet.query={!ex=PulloutSpray_string}PulloutSpray_string:yes&facet.query={!ex=HandlesIncluded_string}HandlesIncluded_string:yes&facet.query={!ex=Electronic_string}Electronic_string:yes&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[0%20TO%201]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[1%20TO%202]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[2%20TO%203]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[4%20TO%205]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[3%20TO%204]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[5%20TO%20*]&facet.query={!ex=ADA_string}ADA_string:yes&facet.query={!ex=WaterSenseCertified_string}WaterSenseCertified_string:yes&facet.query={!ex=WaterfallFaucet_boolean}WaterfallFaucet_boolean:yes&facet.query={!ex=InstallationAvailable_string}InstallationAvailable_string:yes&facet.query={!ex=LowLeadCompliant_string}LowLeadCompliant_string:yes&facet.query={!ex=DrainAssemblyIncluded_string}DrainAssemblyIncluded_string:yes&facet.query={!ex=EscutcheonIncluded_string}EscutcheonIncluded_string:yes&facet.field=NumberOfHandles_numeric&facet.field=pricebook_1_fs&facet.field=SpoutReach_numeric&facet.field=SpoutHeight_numeric&facet.field=FaucetCenters_numeric&facet.field=OverallHeight_numeric&facet.field=FaucetHoles_numeric&facet.field=HandleStyle_string&facet.field=masterFinish_string&facet.field={!ex=manufacturer_string}manufacturer_string&facet.field=HandleMaterial_string&facet.field=ValveType_string&facet.field=Theme_string&facet.field=MountingType_string&qt=/productQuery&qf=sku^9.0%20upc^9.1%20keywords_82_txtws^1.9%20uniqueid^9.0%20series^2.8%20productTitle^1.2%20productid^9.0%20manufacturer^4.0%20masterFinish^1.5%20theme^1.1%20categoryNames_82_txt^0.2%20finish^1.4&pf=keywords_82_txtws^2.1%20productTitle^1.5%20manufacturer^4.0%20finish^1.9&bf=linear(popularity_82_i,1,2)^3.0&q.alt=categories_82_is:108503 > > > Taking out tagging I do not receive the error: > > /solr/productindex/productQuery?fq={!collapse%20field=groupid}&fq=discontinued:false&fq=manufacturer_string:(%22delta%22%20OR%20%22kohler%22)&fq=siteid:82&sort=score%20desc&facet=true&start=0&rows=48&fl=productid,manufacturer,uniqueFinish,uniqueid,productCompositeid,score&facet.query={!ex=onSale}onSale:true&facet.query={!ex=rating}rating:[4%20TO%20*]&facet.query={!ex=rating}rating:[3%20TO%20*]&facet.query={!ex=rating}rating:[2%20TO%20*]&facet.query={!ex=rating}rating:[1%20TO%20*]&facet.query={!ex=MadeinAmerica_boolean}MadeinAmerica_boolean:yes&facet.query={!ex=inStock}inStock:true&facet.query={!ex=PulloutSpray_string}PulloutSpray_string:yes&facet.query={!ex=HandlesIncluded_string}HandlesIncluded_string:yes&facet.query={!ex=Electronic_string}Electronic_string:yes&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[0%20TO%201]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[1%20TO%202]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[2%20TO%203]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[4%20TO%205]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[3%20TO%204]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[5%20TO%20*]&facet.query={!ex=ADA_string}ADA_string:yes&facet.query={!ex=WaterSenseCertified_string}WaterSenseCertified_string:yes&facet.query={!ex=WaterfallFaucet_boolean}WaterfallFaucet_boolean:yes&facet.query={!ex=InstallationAvailable_string}InstallationAvailable_string:yes&facet.query={!ex=LowLeadCompliant_string}LowLeadCompliant_string:yes&facet.query={!ex=DrainAssemblyIncluded_string}DrainAssemblyIncluded_string:yes&facet.query={!ex=EscutcheonIncluded_string}EscutcheonIncluded_string:yes&facet.field=NumberOfHandles_numeric&facet.field=pricebook_1_fs&facet.field=SpoutReach_numeric&f
Re: Error with CollapsingQParserPlugin when trying to use tagging
I've created the following tracker for the issue: https://issues.apache.org/jira/browse/SOLR-5416 -- View this message in context: http://lucene.472066.n3.nabble.com/Error-with-CollapsingQParserPlugin-when-trying-to-use-tagging-tp4098709p4098862.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Highlight: simple.pre/post not being applied always
Could you write down your parameters for highlighting and information about your "analysts_texts" field type? 2013/10/31 Andy Pickler > Solr: 4.5.1 > > I'm sending in a query of "july" and getting back the results and > highlighting I expect with one exception: > > > > > @@@hl@@@Julie@@@endhl@@@ A > > > > > #Month:July > > > > > The simple.pre of @@@hl@@@ and simple.post of @@@endhl@@@ is not being > applied to the one case of the field "#Month:July", even though it's > included in the highlighting section. I've tried changing various > highlighting parameters to no avail. Could someone help me know where to > look for why the pre/post aren't being applied? > > Thanks, > Andy Pickler >
Problem of facet on 170M documents
I have an index with 170M documents, and two of the fields for each doc is "source" and "url". And I want to know the top 500 most frequent urls from Video source. So I did a facet with "fq=source:Video&facet=true&facet.field=url&facet.limit=500", and the matching documents are about 9 millions. The solr cluster is hosted on two ec2 instances each with 4 cpu, and 32G memory. 16G is allocated tfor java heap. 4 master shards on one machine, and 4 replica on another machine. Connected together via zookeeper. Whenever I did the query above, the response is just taking too long and the client will get timed out. Sometimes, when the end user is impatient, so he/she may wait for a few second for the results, and then kill the connection, and then issue the same query again and again. Then the server will have to deal with multiple such heavy queries simultaneously and being so busy that we got "no server hosting shard" error, probably due to lost communication between solr node and zookeeper. Is there any way to deal with such problem? Thanks, Ming