Replication after re adding nodes to cluster (sleeping replicas)

2013-11-01 Thread michael.boom
I have a SolrCloud cluster holding 4 collections, each with with 3 shards and
replication factor = 2.
They all live on 2 machines, and I am currently using this setup for
testing.

However, i would like to connect this test setup to our live application,
just for benchmarking and evaluating if it can handle the big qpm number. 
I am planning also to setup a new machine, and add new nodes manually, one
more replica for each shard on the new machines, in case the first two have
problems handling the big qpm.
But what i would like to do is after I set up the new nodes, to shut down
the new machine and only put it back in the cluster if it's needed.

Thus, getting to the title of this mail: 
After re adding the 3rd machine to the cluster, will the replicas be
automatically synced with the leader, or do i need to manually trigger this
somehow ?

Is there a better idea for having this sleeping  replicas? I bet lots of
people faced this problem, so a best practice must be out there.



-
Thanks,
Michael
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Replication-after-re-adding-nodes-to-cluster-sleeping-replicas-tp4098764.html
Sent from the Solr - User mailing list archive at Nabble.com.


Custom Plugin exception : Plugin init failure for [schema.xml]

2013-11-01 Thread Parvin Gasimzade
Hi all,

I have a problem with custom plugin development in solr 4.x versions. I
have developed custom filter and trying to install it but I got following
exception.

Schema xml :

  




  


Btw, I could not get any exception on 3.x versions. Any help will be
appreciated.

*Exception : *
org.apache.solr.common.SolrException: Plugin init failure for [schema.xml]
fieldType "text_tr": Plugin init failure for [schema.xml] analyzer/filter:
Error instantiating class: 'my_package.TurkishFilterFactory'
at
org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177)
at org.apache.solr.schema.IndexSchema.readSchema(IndexSchema.java:374)
at org.apache.solr.schema.IndexSchema.(IndexSchema.java:122)
at
org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:1018)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:1051)
at org.apache.solr.core.CoreContainer$3.call(CoreContainer.java:634)
at org.apache.solr.core.CoreContainer$3.call(CoreContainer.java:629)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
Caused by: org.apache.solr.common.SolrException: Plugin init failure for
[schema.xml] analyzer/filter: Error instantiating class:
'my_package.TurkishFilterFactory'
at
org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177)
at
org.apache.solr.schema.FieldTypePluginLoader.readAnalyzer(FieldTypePluginLoader.java:392)
at
org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:95)
at
org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:43)
at
org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:151)
... 14 more
Caused by: org.apache.solr.common.SolrException: Error instantiating class:
'my_package.TurkishFilterFactory'
at
org.apache.solr.core.SolrResourceLoader.newInstance(SolrResourceLoader.java:473)
at
org.apache.solr.util.plugin.AbstractPluginLoader.create(AbstractPluginLoader.java:89)
at
org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:151)
... 18 more
Caused by: java.lang.InstantiationException: my_package.TurkishFilterFactory
at java.lang.Class.newInstance(Class.java:359)
at
org.apache.solr.core.SolrResourceLoader.newInstance(SolrResourceLoader.java:470)
... 20 more

Nov 01, 2013 11:59:27 AM org.apache.solr.common.SolrException log
SEVERE: null:org.apache.solr.common.SolrException: Unable to create core:
collection1
at
org.apache.solr.core.CoreContainer.recordAndThrow(CoreContainer.java:1672)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:1057)
at org.apache.solr.core.CoreContainer$3.call(CoreContainer.java:634)
at org.apache.solr.core.CoreContainer$3.call(CoreContainer.java:629)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
Caused by: org.apache.solr.common.SolrException: Plugin init failure for
[schema.xml] fieldType "text_tr": Plugin init failure for [schema.xml]
analyzer/filter: Error instantiating class:
'my_package.TurkishFilterFactory'
at
org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177)
at org.apache.solr.schema.IndexSchema.readSchema(IndexSchema.java:374)
at org.apache.solr.schema.IndexSchema.(IndexSchema.java:122)
at
org.apache.solr.core.CoreContainer.createFromLocal(CoreContainer.java:1018)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:1051)
... 10 more
Caused by: org.apache.solr.common.SolrException: Plugin init failure for
[schema.xml] analyzer/filter: Error instantiating class:
'my_package.TurkishFilterFactory'
at
org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:177)
at
org.apache.solr.schema.FieldTypePluginLoader.readAnalyzer(FieldTypePluginLoader.java:392)
at
org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:95)
at
org.apache.solr.schema.FieldTypePluginLoader.create(FieldTypePluginLoader.java:43)
at
org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.j

Re: Store Solr OpenBitSets In Solr Indexes

2013-11-01 Thread Erick Erickson
Why are you saving this? Because if the bitset you're saving
has anything to do with, say, filter queries, it's probably useless.

The internal bitsets are often based on the internal Lucene doc ID,
which will change when segment merges happen, thus the caution.

Otherwise, theres the binary type you can probably use. It's not very
efficient since I believe it uses base-64 encoding under the covers
though...

Is this an "XY" problem?

Best,
Erick


On Wed, Oct 30, 2013 at 8:06 AM, David Philip
wrote:

> Hi All,
>
> What should be the field type if I have to save solr's open bit set value
> within solr document object and retrieve it later for search?
>
>   OpenBitSet bits = new OpenBitSet();
>
>   bits.set(0);
>   bits.set(1000);
>
>   doc.addField("SolrBitSets", bits);
>
>
> What should be the field type of  SolrBitSets?
>
> Thanks
>


Re: Indexing logs files of thousands of GBs

2013-11-01 Thread Erick Erickson
Throwing a multi-gigabyte file at Solr and expecting it
to index it is asking for a bit too much. You either
have to stream it up and break it apart or something
similar.

And consider what happens if you just index the log as
a single document. How do you search it? Do you return
several G as the result? Most applications break
the log file up into individual documents and index each event
individually to enable searches like
"all OOM errors between 12:00 and 13:00 yesterday" or
similar. How do you expect to do such a thing if it's one
big document?

I may be completely off base here, but I think you need to
define the problem you're solving more clearly. I can flat
guarantee that trying to index a large log file as one document
will be unsatisfactory to search, even if you can get it into
the index.

Best,
Erick


On Wed, Oct 30, 2013 at 12:47 PM, Otis Gospodnetic <
otis.gospodne...@gmail.com> wrote:

> Hi,
>
> Hm, sorry for not helping with this particular issue directly, but it
> looks like you are *uploading* your logs and indexing that way?
> Wouldn't pushing them be a better fit when it comes to log indexing?
> We recently contributed a Logstash output that can index logs to Solr,
> which may be of interest - have a look at
> https://twitter.com/otisg/status/395563043045638144 -- includes a
> little diagram that shows how this fits into the picture.
>
> Otis
> --
> Performance Monitoring * Log Analytics * Search Analytics
> Solr & Elasticsearch Support * http://sematext.com/
>
>
>
> On Wed, Oct 30, 2013 at 9:55 AM, keshari.prerna
>  wrote:
> > Hello,
> >
> > As suggested by Chris, now I am accessing the files using java program
> and
> > creating SolrInputDocument, but i ran into this exception while doing
> > server.add(document). When i tried to increase "ramBufferSizeMB", it
> doesn't
> > let me make it more than 2 gig.
> >
> > org.apache.solr.client.solrj.SolrServerException: Server at
> > http://localhost:8983/solr/logsIndexing returned non ok status:500,
> > message:the request was rejected because its size (2097454) exceeds the
> > configured maximum (2097152)
> > org.apache.commons.fileupload.FileUploadBase$SizeLimitExceededException:
> the
> > request was rejected because its size (2097454) exceeds the configured
> > maximum (2097152)   at
> >
> org.apache.commons.fileupload.FileUploadBase$FileItemIteratorImpl$1.raiseError(FileUploadBase.java:902)
> > at
> >
> org.apache.commons.fileupload.util.LimitedInputStream.checkLimit(LimitedInputStream.java:71)
> > at
> >
> org.apache.commons.fileupload.util.LimitedInputStream.read(LimitedInputStream.java:128)
> > at
> >
> org.apache.commons.fileupload.MultipartStream$ItemInputStream.makeAvailable(MultipartStream.java:977)
> > at
> >
> org.apache.commons.fileupload.MultipartStream$ItemInputStream.read(MultipartStream.java:887)
> > at java.io.InputStream.read(Unknown Source) at
> > org.apache.commons.fileupload.util.Streams.copy(Streams.java:94)
>  at
> > org.apache.commons.fileupload.util.Streams.copy(Streams.java:64)
>  at
> >
> org.apache.commons.fileupload.FileUploadBase.parseRequest(FileUploadBase.java:362)
> > at
> >
> org.apache.commons.fileupload.servlet.ServletFileUpload.parseRequest(ServletFileUpload.java:126)
> > at
> >
> org.apache.solr.servlet.MultipartRequestParser.parseParamsAndFillStreams(SolrRequestParsers.java:344)
> > at
> >
> org.apache.solr.servlet.StandardRequestParser.parseParamsAndFillStreams(SolrRequestParsers.java:397)
> > at
> >
> org.apache.solr.servlet.SolrRequestParsers.parse(SolrRequestParsers.java:115)
> > at
> >
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:244)
> > at
> >
> org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)
> > at
> org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399)
> > at
> >
> org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
> > at
> org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182)
> > at
> org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766)
> > at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450)
> > at org.mortbay.jetty.handler.ContextHand
> > at
> >
> org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:328)
> > at
> >
> org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:211)
> > at
> >
> org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105)
> > at
> org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:121)
> > at
> org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:106)
> > at Filewalker.walk(LogsIndexer.java:48)
> > at Filewalker.main(LogsIndexer.java:69)
> >
> > How do I get rid of this?
> >
> > Thanks,
> > Prerna
> >
> >
> >
> > --
> > View this message in context:
> http://lucene.472066.n3.nabble.com/Indexing-logs-files-of-thousands-of-GBs-

Re: Simple (?) zookeeper question

2013-11-01 Thread Alan Woodward
Unknown document router errors are usually caused by using different solr and 
solrj versions - which version of solr and solrj are you using?

Alan Woodward
www.flax.co.uk


On 1 Nov 2013, at 04:19, Jack Park wrote:

> After digging deeper (slow for a *nix newbee), I uncovered issues with
> the java installation. A step in installation of Oracle Java has it
> that you -install "java" with the path to /bin/java. That done,
> zookeeper seems to be running.
> 
> I booted three cores (on the same box) -- this is the simple one-box
> 3-node cloud test, and used the test code from the Lucidworks course
> to send over and read some documents. That failed with this:
> Unknown document router '{name=compositeId}'
> 
> Lots more research.
> Closer...
> 
> On Thu, Oct 31, 2013 at 5:44 PM, Jack Park  wrote:
>> Latest zookeeper is installed on an Ubuntu server box.
>> Java is 1.7 latest build.
>> whereis points to java just fine.
>> /etc/zookeeper is empty.
>> 
>> boot zookeeper from /bin as sudo ./zkServer.sh start
>> Console says "Started"
>> /etc/zookeeper now has a .pid file
>> In another console, ./zkServer.sh status returns:
>> "It's probably not running"
>> 
>> An interesting fact: the log4j.properties file says there should be a
>> zookeeper.log file in "."; there is no log file. When I do a text
>> search in the zookeeper source code for where it picks up the
>> log4j.properties, nothing is found.
>> 
>> Fascinating, what?  This must be a common beginner's question, not
>> well covered in web-search for my context. Does it ring any bells?
>> 
>> Many thanks.
>> Jack



Re: How to get similarity score between 0 and 1 not relative score

2013-11-01 Thread Peter Keegan
There's another use case for scaling the score. Suppose I want to compute a
custom score based on the weighted sum of:

- product(0.75, relevance score)
- product(0.25, value from another field)

For this to work, both fields must have values between 0-1, for example.
Toby's example using the scale function seems to work, but you have to use
fq to eliminate results with score=0. It seems this is somewhat expensive,
since the scaling can't be done until all results have been collected to
get the max score. Then, are the results resorted? I haven't looked
closely, yet.

Peter


Peter




On Thu, Oct 31, 2013 at 7:48 PM, Toby Lazar  wrote:

> I think you are looking for something like this, though you can omit the fq
> section:
>
>
>
> http://localhost:8983/solr/collection/select?abc=text:bob&q={!func}scale(product(query($abc),1),0,1)&fq={
> !
> frange l=0.9}$q
>
> Also, I don't understand all the fuss about normalized scores.  In the
> linked example, I can see an interest in searching for "apple bannana",
> "zzz yyy xxx qqq kkk ttt rrr 111", etc. and wanting only close matches for
> that point in time.  Would this be a good use for this approach?  I
> understand that the results can change if the documents in the index
> change.
>
> Thanks,
>
> Toby
>
>
>
> On Thu, Oct 31, 2013 at 12:56 AM, Anshum Gupta  >wrote:
>
> > Hi Susheel,
> >
> > Have a look at this:
> > http://wiki.apache.org/lucene-java/ScoresAsPercentages
> >
> > You may really want to reconsider doing that.
> >
> >
> >
> >
> > On Thu, Oct 31, 2013 at 9:41 AM, sushil sharma  > >wrote:
> >
> > > Hi,
> > >
> > > We have a requirement where user would like to see a score (between 0
> to
> > > 1) which can tell how close the input search string is with result
> > string.
> > > So if input was very close but not exact matach, score could be .90
> etc.
> > >
> > > I do understand that we can get score from solr & divide by highest
> score
> > > but that will always show 1 even if we match was not exact.
> > >
> > > Regards,
> > > Susheel
> >
> >
> >
> >
> > --
> >
> > Anshum Gupta
> > http://www.anshumgupta.net
> >
>


Can I combine standardtokenizer with solr.WordDelimiterFilterFactory?

2013-11-01 Thread eShard
Good morning,
Here's the issue: 
I have and ID that consists of two letters and a number.
The whole user title looks like this: Lastname, Firstname (LA12345).
Now, with my current configuration, I can search for LA12345 and find the
user. 
However, when I type in just the number I get zero results.
If I put a wildcard in (*12345) I find the correct record.  
The problem is I changed that user title to use the
worddelimiterfitlerfactory and it seems to work. 
However, I also copy that field into the text field which just uses the
standardtokenizer and I lose the ability to search for 12345 without a
wildcard.
My question is can (or should) I put the worddelimiterfactory in with the
standardtokenizer in the text field?
Or should I just use one or the other?
Thank you,




--
View this message in context: 
http://lucene.472066.n3.nabble.com/Can-I-combine-standardtokenizer-with-solr-WordDelimiterFilterFactory-tp4098814.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: HTTP Basic Authentication with solr's jetty

2013-11-01 Thread Fabiano Sidler
Thus wrote Fabiano Sidler:
> Hi folks!
> 
> I was asking this question last week already on the jetty mailing list, but
> haven't got any answer.
> 
> I'm going to run multiple Solr instances on one server, which arises the need
> of user authentication in front of Solr. I've done the following steps (after
> a lot of others which didn't work):
> 
> === snip ===
> diff -wur solr-4.5.0.orig/example/etc/jetty.xml 
> solr-4.5.0/example/etc/jetty.xml
> --- solr-4.5.0.orig/example/etc/jetty.xml 2013-09-12 14:12:53.0 +0200
> +++ solr-4.5.0/example/etc/jetty.xml 2013-10-28 17:14:49.0 +0100
> @@ -12,6 +12,16 @@
> 
>  
> 
> +
> +  
> +
> +  Test Realm
> +   name="config">/some/path/solr-4.5.0/example/passwords.properties
> +  0
> +
> +  
> +
> +
>  
>  
>  
> diff -wur solr-4.5.0.orig/example/etc/webdefault.xml 
> solr-4.5.0/example/etc/webdefault.xml
> --- solr-4.5.0.orig/example/etc/webdefault.xml2013-09-12 
> 14:12:53.0 +0200
> +++ solr-4.5.0/example/etc/webdefault.xml2013-10-28 
> 17:27:43.0 +0100
> @@ -520,8 +520,14 @@
>/
>TRACE
>  
> -
> +
> +  *
> +
>
> 
> +  
> +BASIC
> +Test Realm
> +  
>  
> diff 
> === snap ===
> 
> The example/password.properties file contains the following test accounts:
> === snip ===
> user: user,user
> admin: admin,admin
> === snap ===
> 
> The problem now is that still no authentication is run by jetty and thus I
> can connect to Solr core without being authenticated. What is still neede
> to finally activate authentication?
> 
> Greetings,
> Fabiano

Hello? Not anyone with an idea or who has already done that?

Greetings,
Fabiano


Re: Can I combine standardtokenizer with solr.WordDelimiterFilterFactory?

2013-11-01 Thread Oussama Jilal

Yes you can and I don't see any reason why you should not

Le 01/11/2013 15:38, eShard a écrit :

Good morning,
Here's the issue:
I have and ID that consists of two letters and a number.
The whole user title looks like this: Lastname, Firstname (LA12345).
Now, with my current configuration, I can search for LA12345 and find the
user.
However, when I type in just the number I get zero results.
If I put a wildcard in (*12345) I find the correct record.
The problem is I changed that user title to use the
worddelimiterfitlerfactory and it seems to work.
However, I also copy that field into the text field which just uses the
standardtokenizer and I lose the ability to search for 12345 without a
wildcard.
My question is can (or should) I put the worddelimiterfactory in with the
standardtokenizer in the text field?
Or should I just use one or the other?
Thank you,




--
View this message in context: 
http://lucene.472066.n3.nabble.com/Can-I-combine-standardtokenizer-with-solr-WordDelimiterFilterFactory-tp4098814.html
Sent from the Solr - User mailing list archive at Nabble.com.




Re: Can I combine standardtokenizer with solr.WordDelimiterFilterFactory?

2013-11-01 Thread Yonik Seeley
On Fri, Nov 1, 2013 at 11:38 AM, eShard  wrote:
> I have and ID that consists of two letters and a number.
> The whole user title looks like this: Lastname, Firstname (LA12345).
> Now, with my current configuration, I can search for LA12345 and find the
> user.
> However, when I type in just the number I get zero results.
> If I put a wildcard in (*12345) I find the correct record.
> The problem is I changed that user title to use the
> worddelimiterfitlerfactory and it seems to work.
> However, I also copy that field into the text field which just uses the
> standardtokenizer and I lose the ability to search for 12345 without a
> wildcard.
> My question is can (or should) I put the worddelimiterfactory in with the
> standardtokenizer in the text field?

copyField worlks before any analysis, so if you want the splitting
behavior on your "text" field, then you'll need to use the
WordDelimiterFilter there also.


-Yonik
http://heliosearch.com -- making solr shine


Re: Simple (?) zookeeper question

2013-11-01 Thread Jack Park
Alan,
That was brilliant!
My test harness was behind a couple of notches.

Hah! So, now we open yet another can of strange looking creatures, namely:

No live SolrServers available to handle this
request:[http://127.0.1.1:8983/solr/collection1]
at 
org.apache.solr.client.solrj.impl.CloudSolrServer.directUpdate(CloudSolrServer.java:347)

3 times, once for each URL I passed into the server. Here is the code:

String zkurl = "10.1.10.178:2181";
String solrurla = "10.1.10.178:8983";
String solrurlb = "10.1.10.178:7574";
String solrurlc = "10.1.10.178:7590";

LBHttpSolrServer sv = new LBHttpSolrServer(solrurla,solrurlb,solrurlc);
CloudSolrServer server = new CloudSolrServer(zkurl,sv);
server.setDefaultCollection("collection1");

I am struggling to imagine how 10.1.10.178 got translated to 127.0.1.1
and the port assignments ignored for each URL passed in.

That error message seems well known to search engines. One suggestion
is to check the zookeeper logs.  According to the zookeeper's log4j
properties, there should be a zookeeper.log in the zookeeper
directory. There is no such log. I went to /etc/zookeeper/Version_2
and looked at log.1 (binary) but could see hints that this might be
where the 127.0.1.1 is coming from: zookeeper sending such an error
message back. This would suggest that, somehow or other, my nodes are
not properly registering themselves, though no error messages were
tossed when each node was booted.

solr.log for node1 only reflects queries from the admin page.

That's what I am working on now.
Thanks!

On Fri, Nov 1, 2013 at 6:03 AM, Alan Woodward  wrote:
> Unknown document router errors are usually caused by using different solr and 
> solrj versions - which version of solr and solrj are you using?
>
> Alan Woodward
> www.flax.co.uk
>
>
> On 1 Nov 2013, at 04:19, Jack Park wrote:
>
>> After digging deeper (slow for a *nix newbee), I uncovered issues with
>> the java installation. A step in installation of Oracle Java has it
>> that you -install "java" with the path to /bin/java. That done,
>> zookeeper seems to be running.
>>
>> I booted three cores (on the same box) -- this is the simple one-box
>> 3-node cloud test, and used the test code from the Lucidworks course
>> to send over and read some documents. That failed with this:
>> Unknown document router '{name=compositeId}'
>>
>> Lots more research.
>> Closer...
>>
>> On Thu, Oct 31, 2013 at 5:44 PM, Jack Park  wrote:
>>> Latest zookeeper is installed on an Ubuntu server box.
>>> Java is 1.7 latest build.
>>> whereis points to java just fine.
>>> /etc/zookeeper is empty.
>>>
>>> boot zookeeper from /bin as sudo ./zkServer.sh start
>>> Console says "Started"
>>> /etc/zookeeper now has a .pid file
>>> In another console, ./zkServer.sh status returns:
>>> "It's probably not running"
>>>
>>> An interesting fact: the log4j.properties file says there should be a
>>> zookeeper.log file in "."; there is no log file. When I do a text
>>> search in the zookeeper source code for where it picks up the
>>> log4j.properties, nothing is found.
>>>
>>> Fascinating, what?  This must be a common beginner's question, not
>>> well covered in web-search for my context. Does it ring any bells?
>>>
>>> Many thanks.
>>> Jack
>


Re: Custom Plugin exception : Plugin init failure for [schema.xml]

2013-11-01 Thread Shawn Heisey
On 11/1/2013 4:18 AM, Parvin Gasimzade wrote:
> I have a problem with custom plugin development in solr 4.x versions. I
> have developed custom filter and trying to install it but I got following
> exception.

Later you indicated that you can use it with Solr 3.x without any problem.

Did you recompile your custom plugin against the Solr jars from the new
version?  There was a *huge* amount of java class refactoring that went
into the 4.0 version as compared to any 3.x version, and that continues
with each new 4.x release.

I would bet that if you tried that recompile, it would fail due to
errors and/or warnings, which you'll need to fix.  There might also be
operational problems that the compiler doesn't find, due to changes in
how the underlying APIs get used.

Thanks,
Shawn



Exclude urls without 'www' from Nutch 1.7 crawl

2013-11-01 Thread Reyes, Mark
I'm currently using Nutch 1.7 to crawl my domain. My issue is specific to URLs 
being indexed as www vs. non-www.

Specifically, after firing the crawl and index to Solr 4.5 then validating the 
results on the front-end with AJAX Solr, the search results page lists 
results/pages that are both 'www' and '' urls such as:

www.mywebsite.com
mywebsite.com
www.mywebsite.com/page1
mywebsite.com/page1

My understanding is that the url filtering (regex-urlfilter.txt) needs 
modification. Are there any regex/nutch experts that could suggest a solution?

Here is the code on paste bin,
http://pastebin.com/Cp6vUxPR

Also on stack overflow,
http://stackoverflow.com/questions/19731904/exclude-urls-without-www-from-nutch-1-7-crawl

Thank you,
Mark


IMPORTANT NOTICE: This e-mail message is intended to be received only by 
persons entitled to receive the confidential information it may contain. E-mail 
messages sent from Bridgepoint Education may contain information that is 
confidential and may be legally privileged. Please do not read, copy, forward 
or store this message unless you are an intended recipient of it. If you 
received this transmission in error, please notify the sender by reply e-mail 
and delete the message and any attachments.

Re: Simple (?) zookeeper question

2013-11-01 Thread Jack Park
/clusterstate.json seems to clearly state that all 3 nodes are alive,
have ranges, and are active.

Still, it would seem that java is still not properly installed.
ZooKeeper is dropping zookeeper.out in the /bin directory, which says
this, among other things:

Server environment:java.home=/usr/local/java/jdk1.7.0_40/jre

Server 
environment:java.class.path=/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../build/classes:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../build/lib/*.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/slf4j-log4j12-1.6.1.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/slf4j-api-1.6.1.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/netty-3.2.2.Final.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/log4j-1.2.15.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/jline-0.9.94.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../zookeeper-3.4.5.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../src/java/lib/*.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../conf:

Server environment:java.library.path=
/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib

There is no /usr/java/...
It's really a mystery where zookeeper is getting these values;
everything else seems right.

But, for me, here's the amazing chunk of traces (cleaned up a bit)

Accepted socket connection from /127.0.0.1:39065
Client attempting to establish new session at /127.0.0.1:39065
Established session 0x1421197e6e90002 with negotiated timeout 15000
for client /127.0.0.1:39065
Got user-level KeeperException when processing
sessionid:0x1421197e6e90002 type:create cxid:0x1 zxid:0xc0 txntype:-1
reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists
for /overseer
Got user-level KeeperException when processing
sessionid:0x1421197e6e90002 type:create cxid:0x3 zxid:0xc1 txntype:-1
reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists
for /overseer
Got user-level KeeperException when processing
sessionid:0x1421197e6e90002 type:delete cxid:0xe zxid:0xc2 txntype:-1
reqpath:n/a Error Path:/live_nodes/127.0.1.1:7590_solr
Error:KeeperErrorCode = NoNode for /live_nodes/127.0.1.1:7590_solr
Got user-level KeeperException when processing
sessionid:0x1421197e6e90002 type:delete cxid:0x9f zxid:0xcd txntype:-1
reqpath:n/a Error Path:/collections/collection1/leaders/shard3
Error:KeeperErrorCode = NoNode for
/collections/collection1/leaders/shard3
2013-10-31 21:01:19,344 [myid:] - INFO  [ProcessThread(sid:0
cport:-1)::PrepRequestProcessor@627] - Got user-level KeeperException
when processing sessionid:0x1421197e6e90002 type:create cxid:0xa0
zxid:0xce txntype:-1 reqpath:n/a Error Path:/overseer
Error:KeeperErrorCode = NodeExists for /overseer
Got user-level KeeperException when processing
sessionid:0x1421197e6e90002 type:create cxid:0xaa zxid:0xd1 txntype:-1
reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists
for /overseer
Accepted socket connection from /10.1.10.180:55528
Client attempting to establish new session at /10.1.10.180:55528
Established session 0x1421197e6e90003 with negotiated timeout 1
for client /10.1.10.180:55528
WARN Exception causing close of session 0x1421197e6e90003 due to
java.io.IOException: Connection reset by peer
Closed socket connection for client /10.1.10.180:55528 which had
sessionid 0x1421197e6e90003

Sockets from 10.1.10.180 are my windoz box shipping solr documents. I
am not sure how I am using 55528 unless that's a solrj behavior.
Connection reset by peer would suggest something in my code, but my
code is a clone of code supplied in a Solr training course. Must be
good. Right?

I also have no clue what is /127.0.0.1:39065 -- that's not one of my nodes.

The quest continues.

On Fri, Nov 1, 2013 at 9:21 AM, Jack Park  wrote:
> Alan,
> That was brilliant!
> My test harness was behind a couple of notches.
>
> Hah! So, now we open yet another can of strange looking creatures, namely:
>
> No live SolrServers available to handle this
> request:[http://127.0.1.1:8983/solr/collection1]
> at 
> org.apache.solr.client.solrj.impl.CloudSolrServer.directUpdate(CloudSolrServer.java:347)
>
> 3 times, once for each URL I passed into the server. Here is the code:
>
> String zkurl = "10.1.10.178:2181";
> String solrurla = "10.1.10.178:8983";
> String solrurlb = "10.1.10.178:7574";
> String solrurlc = "10.1.10.178:7590";
>
> LBHttpSolrServer sv = new LBHttpSolrServer(solrurla,solrurlb,solrurlc);
> CloudSolrServer server = new CloudSolrServer(zkurl,sv);
> server.setDefaultCollection("collection1");
>
> I am struggling to imagine how 10.1.10.178 got translated to 127.0.1.1
> and the port assignments ignored for each URL passed in.
>
> That error message seems well known to search engines. One suggestion
> is to check the zookeeper logs.  According to the zookeeper's log4j
> properties, there should be a zookeeper.log in the zookeeper
> directory.

RE: Exclude urls without 'www' from Nutch 1.7 crawl

2013-11-01 Thread Markus Jelsma
Hi - Use the domain-urlfilter for host, domain and TLD filtering.

Also, please ask questions on the Nutch list, you're on Solr now :)
 
 
-Original message-
> From:Reyes, Mark 
> Sent: Friday 1st November 2013 17:24
> To: solr-user@lucene.apache.org
> Subject: Exclude urls without 'www' from Nutch 1.7 crawl
> 
> I'm currently using Nutch 1.7 to crawl my domain. My issue is specific to 
> URLs being indexed as www vs. non-www.
> 
> Specifically, after firing the crawl and index to Solr 4.5 then validating 
> the results on the front-end with AJAX Solr, the search results page lists 
> results/pages that are both 'www' and '' urls such as:
> 
> www.mywebsite.com
> mywebsite.com
> www.mywebsite.com/page1
> mywebsite.com/page1
> 
> My understanding is that the url filtering (regex-urlfilter.txt) needs 
> modification. Are there any regex/nutch experts that could suggest a solution?
> 
> Here is the code on paste bin,
> http://pastebin.com/Cp6vUxPR
> 
> Also on stack overflow,
> http://stackoverflow.com/questions/19731904/exclude-urls-without-www-from-nutch-1-7-crawl
> 
> Thank you,
> Mark
> 
> 
> IMPORTANT NOTICE: This e-mail message is intended to be received only by 
> persons entitled to receive the confidential information it may contain. 
> E-mail messages sent from Bridgepoint Education may contain information that 
> is confidential and may be legally privileged. Please do not read, copy, 
> forward or store this message unless you are an intended recipient of it. If 
> you received this transmission in error, please notify the sender by reply 
> e-mail and delete the message and any attachments.


Re: Exclude urls without 'www' from Nutch 1.7 crawl

2013-11-01 Thread Furkan KAMACI
As Markus pointed Nutch has a feature for such kind of situation. Here is
Solr list but one more thing for you: www.mywebsite.com and
mywebsite.commay point to "different" pages.


2013/11/1 Markus Jelsma 

> Hi - Use the domain-urlfilter for host, domain and TLD filtering.
>
> Also, please ask questions on the Nutch list, you're on Solr now :)
>
>
> -Original message-
> > From:Reyes, Mark 
> > Sent: Friday 1st November 2013 17:24
> > To: solr-user@lucene.apache.org
> > Subject: Exclude urls without 'www' from Nutch 1.7 crawl
> >
> > I'm currently using Nutch 1.7 to crawl my domain. My issue is specific
> to URLs being indexed as www vs. non-www.
> >
> > Specifically, after firing the crawl and index to Solr 4.5 then
> validating the results on the front-end with AJAX Solr, the search results
> page lists results/pages that are both 'www' and '' urls such as:
> >
> > www.mywebsite.com
> > mywebsite.com
> > www.mywebsite.com/page1
> > mywebsite.com/page1
> >
> > My understanding is that the url filtering (regex-urlfilter.txt) needs
> modification. Are there any regex/nutch experts that could suggest a
> solution?
> >
> > Here is the code on paste bin,
> > http://pastebin.com/Cp6vUxPR
> >
> > Also on stack overflow,
> >
> http://stackoverflow.com/questions/19731904/exclude-urls-without-www-from-nutch-1-7-crawl
> >
> > Thank you,
> > Mark
> >
> >
> > IMPORTANT NOTICE: This e-mail message is intended to be received only by
> persons entitled to receive the confidential information it may contain.
> E-mail messages sent from Bridgepoint Education may contain information
> that is confidential and may be legally privileged. Please do not read,
> copy, forward or store this message unless you are an intended recipient of
> it. If you received this transmission in error, please notify the sender by
> reply e-mail and delete the message and any attachments.
>


Re: Exclude urls without 'www' from Nutch 1.7 crawl

2013-11-01 Thread Reyes, Mark
Noted and will do (that goes twice for the suggestions and putting this on
the nutch list instead).

Thanks all,
Mark



On 11/1/13, 10:53 AM, "Furkan KAMACI"  wrote:

>As Markus pointed Nutch has a feature for such kind of situation. Here is
>Solr list but one more thing for you: www.mywebsite.com and
>mywebsite.commay point to "different" pages.
>
>
>2013/11/1 Markus Jelsma 
>
>> Hi - Use the domain-urlfilter for host, domain and TLD filtering.
>>
>> Also, please ask questions on the Nutch list, you're on Solr now :)
>>
>>
>> -Original message-
>> > From:Reyes, Mark 
>> > Sent: Friday 1st November 2013 17:24
>> > To: solr-user@lucene.apache.org
>> > Subject: Exclude urls without 'www' from Nutch 1.7 crawl
>> >
>> > I'm currently using Nutch 1.7 to crawl my domain. My issue is specific
>> to URLs being indexed as www vs. non-www.
>> >
>> > Specifically, after firing the crawl and index to Solr 4.5 then
>> validating the results on the front-end with AJAX Solr, the search
>>results
>> page lists results/pages that are both 'www' and '' urls such as:
>> >
>> > www.mywebsite.com
>> > mywebsite.com
>> > www.mywebsite.com/page1
>> > mywebsite.com/page1
>> >
>> > My understanding is that the url filtering (regex-urlfilter.txt) needs
>> modification. Are there any regex/nutch experts that could suggest a
>> solution?
>> >
>> > Here is the code on paste bin,
>> > http://pastebin.com/Cp6vUxPR
>> >
>> > Also on stack overflow,
>> >
>> 
>>http://stackoverflow.com/questions/19731904/exclude-urls-without-www-from
>>-nutch-1-7-crawl
>> >
>> > Thank you,
>> > Mark
>> >
>> >
>> > IMPORTANT NOTICE: This e-mail message is intended to be received only
>>by
>> persons entitled to receive the confidential information it may contain.
>> E-mail messages sent from Bridgepoint Education may contain information
>> that is confidential and may be legally privileged. Please do not read,
>> copy, forward or store this message unless you are an intended
>>recipient of
>> it. If you received this transmission in error, please notify the
>>sender by
>> reply e-mail and delete the message and any attachments.
>>


IMPORTANT NOTICE: This e-mail message is intended to be received only by 
persons entitled to receive the confidential information it may contain. E-mail 
messages sent from Bridgepoint Education may contain information that is 
confidential and may be legally privileged. Please do not read, copy, forward 
or store this message unless you are an intended recipient of it. If you 
received this transmission in error, please notify the sender by reply e-mail 
and delete the message and any attachments.

Re: Simple (?) zookeeper question

2013-11-01 Thread Jack Park
The top error message at my test harness is this:

No live SolrServers available to handle this request:
[http://127.0.1.1:8983/solr/collection1,
http://127.0.1.1:7574/solr/collection1,
http://127.0.1.1:7590/solr/collection1]

I have to assume that error message was somehow shipped by zookeeper,
because those servers actually exist, to the test harness, at
10.1.10.178, and if I access any one of them from the browser,
/solr/collection1 does not work, but /solr/#/collection1 does work.

On Fri, Nov 1, 2013 at 10:34 AM, Jack Park  wrote:
> /clusterstate.json seems to clearly state that all 3 nodes are alive,
> have ranges, and are active.
>
> Still, it would seem that java is still not properly installed.
> ZooKeeper is dropping zookeeper.out in the /bin directory, which says
> this, among other things:
>
> Server environment:java.home=/usr/local/java/jdk1.7.0_40/jre
>
> Server 
> environment:java.class.path=/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../build/classes:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../build/lib/*.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/slf4j-log4j12-1.6.1.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/slf4j-api-1.6.1.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/netty-3.2.2.Final.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/log4j-1.2.15.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../lib/jline-0.9.94.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../zookeeper-3.4.5.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../src/java/lib/*.jar:/usr/local/lib/SolrCloud/zookeeper/zookeeper-3.4.5/bin/../conf:
>
> Server environment:java.library.path=
> /usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
>
> There is no /usr/java/...
> It's really a mystery where zookeeper is getting these values;
> everything else seems right.
>
> But, for me, here's the amazing chunk of traces (cleaned up a bit)
>
> Accepted socket connection from /127.0.0.1:39065
> Client attempting to establish new session at /127.0.0.1:39065
> Established session 0x1421197e6e90002 with negotiated timeout 15000
> for client /127.0.0.1:39065
> Got user-level KeeperException when processing
> sessionid:0x1421197e6e90002 type:create cxid:0x1 zxid:0xc0 txntype:-1
> reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists
> for /overseer
> Got user-level KeeperException when processing
> sessionid:0x1421197e6e90002 type:create cxid:0x3 zxid:0xc1 txntype:-1
> reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists
> for /overseer
> Got user-level KeeperException when processing
> sessionid:0x1421197e6e90002 type:delete cxid:0xe zxid:0xc2 txntype:-1
> reqpath:n/a Error Path:/live_nodes/127.0.1.1:7590_solr
> Error:KeeperErrorCode = NoNode for /live_nodes/127.0.1.1:7590_solr
> Got user-level KeeperException when processing
> sessionid:0x1421197e6e90002 type:delete cxid:0x9f zxid:0xcd txntype:-1
> reqpath:n/a Error Path:/collections/collection1/leaders/shard3
> Error:KeeperErrorCode = NoNode for
> /collections/collection1/leaders/shard3
> 2013-10-31 21:01:19,344 [myid:] - INFO  [ProcessThread(sid:0
> cport:-1)::PrepRequestProcessor@627] - Got user-level KeeperException
> when processing sessionid:0x1421197e6e90002 type:create cxid:0xa0
> zxid:0xce txntype:-1 reqpath:n/a Error Path:/overseer
> Error:KeeperErrorCode = NodeExists for /overseer
> Got user-level KeeperException when processing
> sessionid:0x1421197e6e90002 type:create cxid:0xaa zxid:0xd1 txntype:-1
> reqpath:n/a Error Path:/overseer Error:KeeperErrorCode = NodeExists
> for /overseer
> Accepted socket connection from /10.1.10.180:55528
> Client attempting to establish new session at /10.1.10.180:55528
> Established session 0x1421197e6e90003 with negotiated timeout 1
> for client /10.1.10.180:55528
> WARN Exception causing close of session 0x1421197e6e90003 due to
> java.io.IOException: Connection reset by peer
> Closed socket connection for client /10.1.10.180:55528 which had
> sessionid 0x1421197e6e90003
>
> Sockets from 10.1.10.180 are my windoz box shipping solr documents. I
> am not sure how I am using 55528 unless that's a solrj behavior.
> Connection reset by peer would suggest something in my code, but my
> code is a clone of code supplied in a Solr training course. Must be
> good. Right?
>
> I also have no clue what is /127.0.0.1:39065 -- that's not one of my nodes.
>
> The quest continues.
>
> On Fri, Nov 1, 2013 at 9:21 AM, Jack Park  wrote:
>> Alan,
>> That was brilliant!
>> My test harness was behind a couple of notches.
>>
>> Hah! So, now we open yet another can of strange looking creatures, namely:
>>
>> No live SolrServers available to handle this
>> request:[http://127.0.1.1:8983/solr/collection1]
>> at 
>> org.apache.solr.client.solrj.impl.CloudSolrServer.directUpdate(CloudSolrServer.java:347)
>>
>> 3 times, once for each URL I passed into the server. Here is the code:
>>
>> S

Re: Simple (?) zookeeper question

2013-11-01 Thread Shawn Heisey

On 11/1/2013 12:07 PM, Jack Park wrote:

The top error message at my test harness is this:

No live SolrServers available to handle this request:
[http://127.0.1.1:8983/solr/collection1,
http://127.0.1.1:7574/solr/collection1,
http://127.0.1.1:7590/solr/collection1]

I have to assume that error message was somehow shipped by zookeeper,
because those servers actually exist, to the test harness, at
10.1.10.178, and if I access any one of them from the browser,
/solr/collection1 does not work, but /solr/#/collection1 does work.


Those are *base* urls.  By themselves, they return 404. For an example 
of how a base URL is used, try /solr/collection1/select?q=*:* instead.


Any URL with /#/ in it is part of the admin UI, which runs mostly in the 
browser and accesses Solr handlers to gather information. It is not Solr 
itself.


Thanks,
Shawn



Re: Simple (?) zookeeper question

2013-11-01 Thread Jack Park
Thanks. I reviewed clusterstate.json again; those URLs are alive. Why
they are not responding seems to be the mystery du jour.

I reviewed my test suite: it is using field names in schema.xml, and
the server is configured to use the update responders I installed, all
of which work fine in a non-cloud mode.

Thanks
Jack

On Fri, Nov 1, 2013 at 11:12 AM, Shawn Heisey  wrote:
> On 11/1/2013 12:07 PM, Jack Park wrote:
>>
>> The top error message at my test harness is this:
>>
>> No live SolrServers available to handle this request:
>> [http://127.0.1.1:8983/solr/collection1,
>> http://127.0.1.1:7574/solr/collection1,
>> http://127.0.1.1:7590/solr/collection1]
>>
>> I have to assume that error message was somehow shipped by zookeeper,
>> because those servers actually exist, to the test harness, at
>> 10.1.10.178, and if I access any one of them from the browser,
>> /solr/collection1 does not work, but /solr/#/collection1 does work.
>
>
> Those are *base* urls.  By themselves, they return 404. For an example of
> how a base URL is used, try /solr/collection1/select?q=*:* instead.
>
> Any URL with /#/ in it is part of the admin UI, which runs mostly in the
> browser and accesses Solr handlers to gather information. It is not Solr
> itself.
>
> Thanks,
> Shawn
>


Re: Error with CollapsingQParserPlugin when trying to use tagging

2013-11-01 Thread dboychuck
The following unit test in TestCollapseQParserPlugin.java illustrates the
bug

ModifiableSolrParams params = new ModifiableSolrParams();
params.add("q", "*:*");
params.add("fq", "{!collapse field=group_s}");
params.add("defType", "edismax");
params.add("bf", "field(test_ti)");
params.add("fq","{!tag=test_ti}test_ti:5");
params.add("facet","true");
params.add("facet.field","{!ex=test_ti}test_ti");
assertQ(req(params), "*[count(//doc)=1]",
"//doc[./int[@name='test_ti']='5']")


On Thu, Oct 31, 2013 at 6:46 PM, dboychuck [via Lucene] <
ml-node+s472066n4098710...@n3.nabble.com> wrote:

> Here is an example URL that gives the error:
>
> solr/productindex/productQuery?fq={!collapse%20field=groupid}&fq=discontinued:false&fq={!tag=manufacturer_string}manufacturer_string:(%22delta%22%20OR%20%22kohler%22)&fq=siteid:82&sort=score%20desc&facet=true&start=0&rows=48&fl=productid,manufacturer,uniqueFinish,uniqueid,productCompositeid,score&facet.query={!ex=onSale}onSale:true&facet.query={!ex=rating}rating:[4%20TO%20*]&facet.query={!ex=rating}rating:[3%20TO%20*]&facet.query={!ex=rating}rating:[2%20TO%20*]&facet.query={!ex=rating}rating:[1%20TO%20*]&facet.query={!ex=MadeinAmerica_boolean}MadeinAmerica_boolean:yes&facet.query={!ex=inStock}inStock:true&facet.query={!ex=PulloutSpray_string}PulloutSpray_string:yes&facet.query={!ex=HandlesIncluded_string}HandlesIncluded_string:yes&facet.query={!ex=Electronic_string}Electronic_string:yes&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[0%20TO%201]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[1%20TO%202]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[2%20TO%203]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[4%20TO%205]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[3%20TO%204]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[5%20TO%20*]&facet.query={!ex=ADA_string}ADA_string:yes&facet.query={!ex=WaterSenseCertified_string}WaterSenseCertified_string:yes&facet.query={!ex=WaterfallFaucet_boolean}WaterfallFaucet_boolean:yes&facet.query={!ex=InstallationAvailable_string}InstallationAvailable_string:yes&facet.query={!ex=LowLeadCompliant_string}LowLeadCompliant_string:yes&facet.query={!ex=DrainAssemblyIncluded_string}DrainAssemblyIncluded_string:yes&facet.query={!ex=EscutcheonIncluded_string}EscutcheonIncluded_string:yes&facet.field=NumberOfHandles_numeric&facet.field=pricebook_1_fs&facet.field=SpoutReach_numeric&facet.field=SpoutHeight_numeric&facet.field=FaucetCenters_numeric&facet.field=OverallHeight_numeric&facet.field=FaucetHoles_numeric&facet.field=HandleStyle_string&facet.field=masterFinish_string&facet.field={!ex=manufacturer_string}manufacturer_string&facet.field=HandleMaterial_string&facet.field=ValveType_string&facet.field=Theme_string&facet.field=MountingType_string&qt=/productQuery&qf=sku^9.0%20upc^9.1%20keywords_82_txtws^1.9%20uniqueid^9.0%20series^2.8%20productTitle^1.2%20productid^9.0%20manufacturer^4.0%20masterFinish^1.5%20theme^1.1%20categoryNames_82_txt^0.2%20finish^1.4&pf=keywords_82_txtws^2.1%20productTitle^1.5%20manufacturer^4.0%20finish^1.9&bf=linear(popularity_82_i,1,2)^3.0&q.alt=categories_82_is:108503
>
>
> Taking out tagging I do not receive the error:
>
> /solr/productindex/productQuery?fq={!collapse%20field=groupid}&fq=discontinued:false&fq=manufacturer_string:(%22delta%22%20OR%20%22kohler%22)&fq=siteid:82&sort=score%20desc&facet=true&start=0&rows=48&fl=productid,manufacturer,uniqueFinish,uniqueid,productCompositeid,score&facet.query={!ex=onSale}onSale:true&facet.query={!ex=rating}rating:[4%20TO%20*]&facet.query={!ex=rating}rating:[3%20TO%20*]&facet.query={!ex=rating}rating:[2%20TO%20*]&facet.query={!ex=rating}rating:[1%20TO%20*]&facet.query={!ex=MadeinAmerica_boolean}MadeinAmerica_boolean:yes&facet.query={!ex=inStock}inStock:true&facet.query={!ex=PulloutSpray_string}PulloutSpray_string:yes&facet.query={!ex=HandlesIncluded_string}HandlesIncluded_string:yes&facet.query={!ex=Electronic_string}Electronic_string:yes&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[0%20TO%201]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[1%20TO%202]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[2%20TO%203]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[4%20TO%205]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[3%20TO%204]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[5%20TO%20*]&facet.query={!ex=ADA_string}ADA_string:yes&facet.query={!ex=WaterSenseCertified_string}WaterSenseCertified_string:yes&facet.query={!ex=WaterfallFaucet_boolean}WaterfallFaucet_boolean:yes&facet.query={!ex=InstallationAvailable_string}InstallationAvailable_string:yes&facet.query={!ex=LowLeadCompliant_string}LowLeadCompliant_string:yes&facet.query={!ex=DrainAssemblyIncluded_string}DrainAssemblyIncluded_string:yes&facet.query={!ex=EscutcheonIncluded_string}EscutcheonIncluded_string:yes&facet.field=NumberOfHandles_numeric&facet.field=pricebook_1_fs&facet.field=SpoutReach_numeric&f

Re: Error with CollapsingQParserPlugin when trying to use tagging

2013-11-01 Thread dboychuck
I've created the following tracker for the issue:
https://issues.apache.org/jira/browse/SOLR-5416



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Error-with-CollapsingQParserPlugin-when-trying-to-use-tagging-tp4098709p4098862.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Highlight: simple.pre/post not being applied always

2013-11-01 Thread Furkan KAMACI
Could you write down your parameters for highlighting and information about
your "analysts_texts" field type?


2013/10/31 Andy Pickler 

> Solr: 4.5.1
>
> I'm sending in a query of "july" and getting back the results and
> highlighting I expect with one exception:
>
> 
> 
> 
> @@@hl@@@Julie@@@endhl@@@ A
> 
> 
> 
> 
> #Month:July
> 
> 
> 
>
> The simple.pre of @@@hl@@@ and simple.post of @@@endhl@@@ is not being
> applied to the one case of the field "#Month:July", even though it's
> included in the highlighting section.  I've tried changing various
> highlighting parameters to no avail.  Could someone help me know where to
> look for why the pre/post aren't being applied?
>
> Thanks,
> Andy Pickler
>


Problem of facet on 170M documents

2013-11-01 Thread Mingfeng Yang
I have an index with 170M documents, and two of the fields for each doc is
"source" and "url".  And I want to know the top 500 most frequent urls from
Video source.

So I did a facet with
 "fq=source:Video&facet=true&facet.field=url&facet.limit=500", and the
matching documents are about 9 millions.

The solr cluster is hosted on two ec2 instances each with 4 cpu, and  32G
memory. 16G is allocated tfor java heap.  4 master shards on one machine,
and 4 replica on another machine. Connected together via zookeeper.

Whenever I did the query above, the response is just taking too long and
the client will get timed out. Sometimes,  when the end user is impatient,
so he/she may wait for a few second for the results, and then kill the
connection, and then issue the same query again and again.  Then the server
will have to deal with multiple such heavy queries simultaneously and
 being so busy that we got "no server hosting shard" error, probably due to
lost communication between solr node and zookeeper.

Is there any way to deal with such problem?

Thanks,
Ming