Re: AJAX Solr returning the default wildcard *:* and not what I query

2013-10-31 Thread Raymond Wiker
The parameters indicate a jQuery.ajax call with result type "jsonp" - a

Re: Unable to add mahout classifier

2013-10-31 Thread Koji Sekiguchi

Caused by: java.lang.ClassCastException: class 
com.mahout.solr.classifier.CategorizeDocumentFactory
 at java.lang.Class.asSubclass(Unknown Source)
 at 
org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:433)
 at 
org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:381)
 at org.apache.solr.core.SolrCore.createInstance(SolrCore.java:526)
 ... 21 more


There seems to be a problem related class loaders, e.g. 
CategorizeDocumentFactory
which extends UpdateRequestProcessorFactory, loaded by class loader "B",
but Solr core has loaded UpdateRequestProcessorFactory via class loader "A"
or something like that...

koji
--
http://www.rondhuit.com/


Re: Phrase query combined with term query for maximum accuracy

2013-10-31 Thread michael.boom
Thanks Jack!
I tried it and i get a really funny behaviour: I have two collections,
having the same solrconfig.xml and the same schema definition, except for
the type of some fields, which in collection_DE are customized for German
languange and in collection_US for English

  
 
  
  
  
  
  

  

  
 
  
  
  

  

Fields "title" and "text" have the corresponding type (text_de in
collection_DE and text_en in collection_US)

Now, when i run this query:
/solr/collection_US/select/?q=title:"blue hat"^100 OR text:"blue hat"^50 OR
title:(blue hat)^30 OR text:(blue
hat)^10&fq=active:true&start=0&rows=40&sort=score+desc&fl=*,score&country=US

i get error:

No live SolrServers available to handle this
request:[http://xxx:8983/solr/collection_US_shard2_replica1,
http://xxx:8983/solr/collection_US_shard2_replica2]","trace":"org.apache.solr.common.SolrException:
org.apache.solr.client.solrj.SolrServerException: No live SolrServers
available to handle this
request:[http://xx:8983/solr/collection_US_shard2_replica1,
http://xx:8983/solr/collection_US_shard2_replica2]\n\tat
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:302)\n\tat
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)\n\tat
org.apache.solr.core.SolrCore.execute(SolrCore.java:1904)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:659)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:362)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:158)\n\tat
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1489)\n\tat
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:517)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:138)\n\tat
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:540)\n\tat
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:213)\n\tat
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1097)\n\tat
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:446)\n\tat
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:175)\n\tat
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1031)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:136)\n\tat
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:200)\n\tat
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:109)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)\n\tat
org.eclipse.jetty.rewrite.handler.RewriteHandler.handle(RewriteHandler.java:317)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)\n\tat
org.eclipse.jetty.server.Server.handle(Server.java:445)\n\tat
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:269)\n\tat
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:229)\n\tat
org.eclipse.jetty.io.AbstractConnection$ReadCallback.run(AbstractConnection.java:358)\n\tat
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:601)\n\tat
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:532)\n\tat
java.lang.Thread.run(Thread.java:724)\nCaused by:
org.apache.solr.client.solrj.SolrServerException: No live SolrServers
available to handle this
request:[http://xxx:8983/solr/collection_US_shard2_replica1,
http://xxx:8983/solr/collection_US_shard2_replica2]\n\tat
org.apache.solr.client.solrj.impl.LBHttpSolrServer.request(LBHttpSolrServer.java:333)\n\tat
org.apache.solr.handler.component.HttpShardHandlerFactory.makeLoadBalancedRequest(HttpShardHandlerFactory.java:214)\n\tat
org.apache.solr.handler.component.HttpShardHandler$1.call(HttpShardHandler.java:158)\n\tat
org.apache.solr.handler.component.HttpShardHandler$1.call(HttpShardHandler.java:119)\n\tat
java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)\n\tat
java.util.concurrent.FutureTask.run(FutureTask.java:166)\n\tat
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)\n\tat
java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)\n\tat
java.util.concurrent.FutureTask.run(FutureTask.java:166)\n\tat
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)\n\tat
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)\n\t...
1 more\nCaused by:
org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: Server
at http://xxx:8983/solr/collection_US_shard2_replica2 returned non ok
status:500, message:Server Error\n\tat
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:385)\n\tat
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)\n\tat
org

Re: Phrase query combined with term query for maximum accuracy

2013-10-31 Thread michael.boom
One more thing i just noticed:
if for collection_US i try to search for 
title:"blue hat"^100 OR text:"blue hat"^50-> i get the same error
but if i search for :
title:"blue hat"^100 OR text:"bluehat"^50 -> it works fine



-
Thanks,
Michael
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Phrase-query-combined-with-term-query-for-maximum-accuracy-tp4098215p4098599.html
Sent from the Solr - User mailing list archive at Nabble.com.


RE: solr 4.5.0 configuration Error: org.apache.solr.common.SolrException:org.apache.solr.common.SolrException: Could not load config file .../solrconfig.xml

2013-10-31 Thread Elena Camossi
Thanks Shawn,

> > Error loading class 'solr.clustering.ClusteringComponent'
> >
> > Please check your logs for more information Log4j
> > (org.slf4j.impl.Log4jLoggerFactory)
> > TimeLevel   Logger  Message
> > 17:36:43WARNSolrResourceLoader  Can't find (or read)
> > directory to add to classloader: ../../../contrib/extraction/lib
> > (resolved
> > as: /usr/share/solr/ckan/../../../contrib/extraction/lib).
> 
> Your solrconfig.xml file includes the ClusteringComponent, but you don't
> have the jars required for that component available.  Your solrconfig file
does
> have a bunch of  directives, but they don't point anywhere that's
valid --
> they assume that the entire Solr download is available, not just what's in
the
> example dir.  The jar for that particular component can be found in the
> download as dist/solr-clustering-X.X.X.jar ... but it is likely to also
require
> additional jars, such as those found in contrib/clustering/lib.
> 
> When it comes to extra jars for contrib or third-party components, the
best
> thing to do is remove all  directives from solrconfig.xml and put the
jars
> in ${solr.solr.home}/lib.  For you that location would be
/usr/share/solr/lib.
> Solr automatically looks in this location without any extra configuration.
>

I have put all the jar in the ${solr.solr.home}/lib folder, and now Solr
starts correctly!
There are still warnings, but apparently Solr works

11:44:37WARNSolrResourceLoader  Can't find (or read)
directory to add to classloader: ../../../contrib/extraction/lib (resolved
as: /usr/share/solr/ckan/../../../contrib/extraction/lib).
11:44:37WARNSolrResourceLoader  Can't find (or read)
directory to add to classloader: ../../../dist/ (resolved as:
/usr/share/solr/ckan/../../../dist).
11:44:37WARNSolrResourceLoader  Can't find (or read)
directory to add to classloader: ../../../contrib/clustering/lib/ (resolved
as: /usr/share/solr/ckan/../../../contrib/clustering/lib).
11:44:37WARNSolrResourceLoader  Can't find (or read)
directory to add to classloader: ../../../dist/ (resolved as:
/usr/share/solr/ckan/../../../dist).
11:44:37WARNSolrResourceLoader  Can't find (or read)
directory to add to classloader: ../../../contrib/langid/lib/ (resolved as:
/usr/share/solr/ckan/../../../contrib/langid/lib).
11:44:37WARNSolrResourceLoader  Can't find (or read)
directory to add to classloader: ../../../dist/ (resolved as:
/usr/share/solr/ckan/../../../dist).
11:44:37WARNSolrResourceLoader  Can't find (or read)
directory to add to classloader: ../../../contrib/velocity/lib (resolved as:
/usr/share/solr/ckan/../../../contrib/velocity/lib).
11:44:37WARNSolrResourceLoader  Can't find (or read)
directory to add to classloader: ../../../dist/ (resolved as:
/usr/share/solr/ckan/../../../dist).

 
 
> Further advice - remove things you don't need from your config.  If you're
> not planning to use the clustering component, take it out.  Also remove
any
> handlers that refer to components you won't be using -- the /browse
> handler is a prime example of something that most people don't need.

Thanks for this tip. 
The original CKAN installation instructions for Solr were for Solr 1.4.1, so
most likely with that version solrconf.xml was simpler and no editig was
required.
I will send the CKAN developers ml the new instructions for configuring Solr
4.5, asking them which Solr components are actually required by CKAN, so
solrconf.xml can be changed accordingly.

Thanks a lot for your precious help!

Regards,
-Elena




Query OR operator triggers weird exception

2013-10-31 Thread michael.boom
I run a set of queries using the AdminUI and some of them trigger a weird
error:
"error": {
"msg": "org.apache.solr.client.solrj.SolrServerException: No live
SolrServers available to handle this request:...
"code": 500
  }
Here's the pattern, using the edismax parser:
title:"blue hat" OR text:"blue hat"   -->error above
title:"blue hat" OR text:"bluehat"--> OK
title:"blue hat" OR text:(blue hat)   --> OK  
title:(blue hat) OR text:(blue hat)--> OK   

Any idea what is wrong here? 
Thanks!



-
Thanks,
Michael
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Query-OR-operator-triggers-weird-exception-tp4098605.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Query OR operator triggers weird exception

2013-10-31 Thread Jack Krupansky
Maybe your schema has "omitted" position information for the "text" field 
which is required to query phrases. That would throw an exception on every 
server for a distributed query. Still, that does seem like an odd exception 
for a simple query problem.


-- Jack Krupansky

-Original Message- 
From: michael.boom

Sent: Thursday, October 31, 2013 7:31 AM
To: solr-user@lucene.apache.org
Subject: Query OR operator triggers weird exception

I run a set of queries using the AdminUI and some of them trigger a weird
error:
"error": {
   "msg": "org.apache.solr.client.solrj.SolrServerException: No live
SolrServers available to handle this request:...
   "code": 500
 }
Here's the pattern, using the edismax parser:
title:"blue hat" OR text:"blue hat"   -->error above
title:"blue hat" OR text:"bluehat"--> OK
title:"blue hat" OR text:(blue hat)   --> OK
title:(blue hat) OR text:(blue hat)--> OK

Any idea what is wrong here?
Thanks!



-
Thanks,
Michael
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Query-OR-operator-triggers-weird-exception-tp4098605.html
Sent from the Solr - User mailing list archive at Nabble.com. 



Re: Query OR operator triggers weird exception

2013-10-31 Thread michael.boom
Thanks Jack!
Some more info: I looked a little bit and tried the problem query,
undistributed, on each shard:
shard2_replica1 and shard2_replica2 throw this error:
"responseHeader":{
"status":500,
"QTime":2,
"params":{
  "lowercaseOperators":"true",
  "indent":"true",
  "q":"title:\"red shoes\" OR text:\"red shoe\"",
  "distrib":"false",
  "stopwords":"true",
  "wt":"json",
  "defType":"edismax"}},
  "error":{
"trace":"java.lang.ArrayIndexOutOfBoundsException\n",
"code":500}}

On other shards the query works fine.



-
Thanks,
Michael
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Query-OR-operator-triggers-weird-exception-tp4098605p4098607.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Query OR operator triggers weird exception

2013-10-31 Thread michael.boom
I also narrowed my problem to the text field.
simple query :   title:"red shoes"   works
but text:"red shoes"   does not.

Could you extend a little bit how could my schema omitted position
information?
I'm not really sure what you mean by that.

Thank you!



-
Thanks,
Michael
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Query-OR-operator-triggers-weird-exception-tp4098605p4098609.html
Sent from the Solr - User mailing list archive at Nabble.com.


safe to delete old index

2013-10-31 Thread Christopher Gross
I have Solr 4.1 running in the SolrCloud mode.  My largest collection has 2
index directories (and an index.properties & replication.properties in that
directory).  Is it safe to remove the older index not listed in
index.properties?  I'm running low on disk space, otherwise I'd have just
left it alone.

Thanks!

-- Chris


Re: Query OR operator triggers weird exception

2013-10-31 Thread michael.boom
After restarting my servers this was the first error i got when trying to
make the same query:

 {
  "responseHeader":{
"status":500,
"QTime":336,
"params":{
  "lowercaseOperators":"true",
  "indent":"true",
  "q":"text:\"blue cat\"",
  "distrib":"false",
  "stopwords":"true",
  "wt":"json",
  "defType":"edismax"}},
  "error":{
"msg":"-103",
"trace":"java.lang.ArrayIndexOutOfBoundsException: -103\n\tat
org.apache.lucene.codecs.lucene41.ForUtil.skipBlock(ForUtil.java:219)\n\tat
org.apache.lucene.codecs.lucene41.Lucene41PostingsReader$BlockDocsAndPositionsEnum.skipPositions(Lucene41PostingsReader.java:958)\n\tat
org.apache.lucene.codecs.lucene41.Lucene41PostingsReader$BlockDocsAndPositionsEnum.nextPosition(Lucene41PostingsReader.java:988)\n\tat
org.apache.lucene.search.ExactPhraseScorer.phraseFreq(ExactPhraseScorer.java:213)\n\tat
org.apache.lucene.search.ExactPhraseScorer.nextDoc(ExactPhraseScorer.java:134)\n\tat
org.apache.lucene.search.Scorer.score(Scorer.java:64)\n\tat
org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:624)\n\tat
org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:297)\n\tat
org.apache.solr.search.SolrIndexSearcher.getDocListNC(SolrIndexSearcher.java:1494)\n\tat
org.apache.solr.search.SolrIndexSearcher.getDocListC(SolrIndexSearcher.java:1363)\n\tat
org.apache.solr.search.SolrIndexSearcher.search(SolrIndexSearcher.java:474)\n\tat
org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:434)\n\tat
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:208)\n\tat
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)\n\tat
org.apache.solr.core.SolrCore.execute(SolrCore.java:1904)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:659)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:362)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:158)\n\tat
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1489)\n\tat
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:517)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:138)\n\tat
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:540)\n\tat
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:213)\n\tat
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1097)\n\tat
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:446)\n\tat
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:175)\n\tat
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1031)\n\tat
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:136)\n\tat
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:200)\n\tat
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:109)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)\n\tat
org.eclipse.jetty.rewrite.handler.RewriteHandler.handle(RewriteHandler.java:317)\n\tat
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)\n\tat
org.eclipse.jetty.server.Server.handle(Server.java:445)\n\tat
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:269)\n\tat
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:229)\n\tat
org.eclipse.jetty.io.AbstractConnection$ReadCallback.run(AbstractConnection.java:358)\n\tat
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:601)\n\tat
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:532)\n\tat
java.lang.Thread.run(Thread.java:724)\n",
"code":500}}



-
Thanks,
Michael
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Query-OR-operator-triggers-weird-exception-tp4098605p4098614.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: ReplicationHandler - SnapPull failed to download a file completely.

2013-10-31 Thread Shalom Ben-Zvii Kazaz
Shawn, Thank you for your answer.
for the purpose of testing it we have a test environment where we are not
indexing anymore. We also disabled the DIH delta import. so as I understand
there shouldn't be any commits on the master.
I also tried with
50:50:50
and get the same failure.

I tried changing and increasing various parameters on the master and slave
but no luck yet.
the master is functioning ok, we do have search results so I assume there
is no index corruption on the master side.
just to mention , we have done that many times before in the past few
years, this started just now when we upgraded our solr from version 3.6 to
version 4.3 and we reindexed all documents.

if we have no solution soon, and this is holding an upgrade to our
production site and various customers, do you think we can copy the index
directory from the master to the slave and hope that future replication
will work ?

Thank you again.

Shalom





On Wed, Oct 30, 2013 at 10:00 PM, Shawn Heisey  wrote:

> On 10/30/2013 1:49 PM, Shalom Ben-Zvi Kazaz wrote:
>
>> we are continuously getting this exception during replication from
>> master to slave. our index size is 9.27 G and we are trying to replicate
>> a slave from scratch.
>> Its a different file each time , sometimes we get to 60% replication
>> before it fails and sometimes only 10%, we never managed a successful
>> replication.
>>
>
> 
>
>
>  this is the master setup:
>>
>> |
>> 
>>   commit
>>   startup<**/str>
>>   stopwords.**txt,spellings.txt,synonyms.**
>> txt,protwords.txt,elevate.xml,**currency.xml
>>   **00:00:50
>> 
>> 
>>
>
> I assume that you're probably doing commits fairly often, resulting in a
> lot of merge activity that frequently deletes segments.  That
> "commitReserveDuration" parameter needs to be made larger.  I would imagine
> that it takes a lot more than 50 seconds to do the replication - even if
> you've got an extremely fast network, replicating 9.7GB probably takes
> several minutes.
>
> From the wiki page on replication:  "If your commits are very frequent and
> network is particularly slow, you can tweak an extra attribute  name="commitReserveDuration">**00:00:10. This is roughly the time
> taken to download 5MB from master to slave. Default is 10 secs."
>
> http://wiki.apache.org/solr/**SolrReplication#Master
>
> You've said that your network is not slow, but with that much data, all
> networks are slow.
>
> Thanks,
> Shawn
>
>


Highlight: simple.pre/post not being applied always

2013-10-31 Thread Andy Pickler
Solr: 4.5.1

I'm sending in a query of "july" and getting back the results and
highlighting I expect with one exception:




@@@hl@@@Julie@@@endhl@@@ A




#Month:July




The simple.pre of @@@hl@@@ and simple.post of @@@endhl@@@ is not being
applied to the one case of the field "#Month:July", even though it's
included in the highlighting section.  I've tried changing various
highlighting parameters to no avail.  Could someone help me know where to
look for why the pre/post aren't being applied?

Thanks,
Andy Pickler


Re: safe to delete old index

2013-10-31 Thread Greg Walters
Hi Chris,

In my experience it is safe to delete older indexes like that. You might want 
to check if the index is in use prior to deleting it via the `lsof` command on 
linux or the equivalent on other platforms. I've found that most times, if the 
index isn't the one specified in index.properties then `lsof` won't show Solr 
as using it.

FWIW I'm pretty sure there's a bug in Jira about old indexes not getting purged 
but I can't find it right now.

Thanks,
Greg

On 2013Oct 31,, at 7:32 AM, Christopher Gross  wrote:

> I have Solr 4.1 running in the SolrCloud mode.  My largest collection has 2
> index directories (and an index.properties & replication.properties in that
> directory).  Is it safe to remove the older index not listed in
> index.properties?  I'm running low on disk space, otherwise I'd have just
> left it alone.
> 
> Thanks!
> 
> -- Chris



Re: safe to delete old index

2013-10-31 Thread Greg Walters
* You might want to check that the index is NOT in use *

(It's still early and dark here!)

Greg

On 2013Oct 31,, at 9:57 AM, Greg Walters  wrote:

> Hi Chris,
> 
> In my experience it is safe to delete older indexes like that. You might want 
> to check if the index is in use prior to deleting it via the `lsof` command 
> on linux or the equivalent on other platforms. I've found that most times, if 
> the index isn't the one specified in index.properties then `lsof` won't show 
> Solr as using it.
> 
> FWIW I'm pretty sure there's a bug in Jira about old indexes not getting 
> purged but I can't find it right now.
> 
> Thanks,
> Greg
> 
> On 2013Oct 31,, at 7:32 AM, Christopher Gross  wrote:
> 
>> I have Solr 4.1 running in the SolrCloud mode.  My largest collection has 2
>> index directories (and an index.properties & replication.properties in that
>> directory).  Is it safe to remove the older index not listed in
>> index.properties?  I'm running low on disk space, otherwise I'd have just
>> left it alone.
>> 
>> Thanks!
>> 
>> -- Chris
> 



Re: AJAX Solr returning the default wildcard *:* and not what I query

2013-10-31 Thread Reyes, Mark
I just tweaked the reuters.js example to listen to the window.location
object and it resolved the wildcard returns.

I put it on pastebin,
http://pastebin.com/GyC4RMva


Thanks for the reply everyone,
Mark
---


P. 866.475.0317 x 3244
Bridgepoint Education
INNOVATIVE SOLUTIONS THAT ADVANCE LEARNING SM




On 10/31/13, 12:23 AM, "Raymond Wiker"  wrote:

>The parameters indicate a jQuery.ajax call with result type "jsonp" - a
>

Re: [SolrCloud-Solrj] Document router problem connecting to Zookeeper ensemble

2013-10-31 Thread Alessandro Benedetti
You are right man !
The version of Solr 4.5, the version of SolrJ is 4.4 .
No the error seems to have solved, i have other problems, but i let you
know.

i have been so stupid, thank you very much !



2013/10/30 Anshum Gupta 

> Hi Alessandro,
>
> What version of Solr are you running and what's the version of SolrJ? I am
> guessing they are different.
>
>
>
>
> On Wed, Oct 30, 2013 at 8:32 PM, Alessandro Benedetti <
> benedetti.ale...@gmail.com> wrote:
>
> > I have a zookeeper ensemble hotes in one amazon server.
> > Using the CloudSolrServer and trying to connect , I obtain this nreally
> > unusual error :
> >
> > 969 [main] INFO org.apache.solr.common.cloud.ConnectionManager - Client
> is
> > connected to ZooKeeper
> > 1043 [main] INFO org.apache.solr.common.cloud.ZkStateReader - Updating
> > cluster state from ZooKeeper...
> > Exception in thread "main" org.apache.solr.common.SolrException: Unknown
> > document router '{name=implicit}'
> > at org.apache.solr.common.cloud.DocRouter.getDocRouter(DocRouter.java:46)
> >
> > Although in my collection I have the compositeId strategy for routing (
> > from the clusterState.json ) .
> >
> > This is how I instantiate the server :
> >
> > CloudSolrServer server;
> > server = new CloudSolrServer("
> > ec2-xx.xx.xx.eu-west-1.compute.amazonaws.com:2181,
> > ec2-xx.xx.xx.eu-west-1.compute.amazonaws.com:2182,
> > ec2-xx.xx.xx.eu-west-1.compute.amazonaws.com:2183");
> > server.setDefaultCollection("example");
> > SolrPingResponse ping = server.ping();
> >
> > Any hint ?
> > --
> > --
> >
> > Benedetti Alessandro
> > Visiting card : http://about.me/alessandro_benedetti
> >
> > "Tyger, tyger burning bright
> > In the forests of the night,
> > What immortal hand or eye
> > Could frame thy fearful symmetry?"
> >
> > William Blake - Songs of Experience -1794 England
> >
>
>
>
> --
>
> Anshum Gupta
> http://www.anshumgupta.net
>



-- 
--

Benedetti Alessandro
Visiting card : http://about.me/alessandro_benedetti

"Tyger, tyger burning bright
In the forests of the night,
What immortal hand or eye
Could frame thy fearful symmetry?"

William Blake - Songs of Experience -1794 England


Re: safe to delete old index

2013-10-31 Thread Christopher Gross
It's all good.  I actually read it "right" the first time. :)

Thanks Greg!

-- Chris


On Thu, Oct 31, 2013 at 10:58 AM, Greg Walters wrote:

> * You might want to check that the index is NOT in use *
>
> (It's still early and dark here!)
>
> Greg
>
> On 2013Oct 31,, at 9:57 AM, Greg Walters  wrote:
>
> > Hi Chris,
> >
> > In my experience it is safe to delete older indexes like that. You might
> want to check if the index is in use prior to deleting it via the `lsof`
> command on linux or the equivalent on other platforms. I've found that most
> times, if the index isn't the one specified in index.properties then `lsof`
> won't show Solr as using it.
> >
> > FWIW I'm pretty sure there's a bug in Jira about old indexes not getting
> purged but I can't find it right now.
> >
> > Thanks,
> > Greg
> >
> > On 2013Oct 31,, at 7:32 AM, Christopher Gross  wrote:
> >
> >> I have Solr 4.1 running in the SolrCloud mode.  My largest collection
> has 2
> >> index directories (and an index.properties & replication.properties in
> that
> >> directory).  Is it safe to remove the older index not listed in
> >> index.properties?  I'm running low on disk space, otherwise I'd have
> just
> >> left it alone.
> >>
> >> Thanks!
> >>
> >> -- Chris
> >
>
>


Re: ReplicationHandler - SnapPull failed to download a file completely.

2013-10-31 Thread Shawn Heisey
On 10/31/2013 7:26 AM, Shalom Ben-Zvii Kazaz wrote:
> Shawn, Thank you for your answer.
> for the purpose of testing it we have a test environment where we are not
> indexing anymore. We also disabled the DIH delta import. so as I understand
> there shouldn't be any commits on the master.
> I also tried with
> 50:50:50
> and get the same failure.

If it's in an environment where there are no commits, that's really
odd.  I would suspect underlying filesystem or network issues.  There's
one problem that's not well known, but is very common - problems with
NIC firmware, most commonly Broadcom NICs.  These problems result in
things working correctly almost all the time, but when there is a high
network load, things break in strange ways, and the resulting errors
rarely look like they are network-related.

Most embedded NICs are either Broadcom or Realtek, both of which are
famous for their firmware problems.  Broadcom NICs are very common on
Dell and HP servers.  Upgrading the firmware (which is not usually the
same thing as upgrading the driver) is the only fix.  NICs from other
manufacturers also have upgradable firmware, but don't usually have the
same kind of high-profile problems as Broadcom.

The NIC firmware might not have anything to do with this problem, but
it's the only thing left that I can think of.  I personally haven't used
replication since Solr 1.4.1, but a lot of people do.  I can't say that
there's no bugs, but so far I'm not seeing the kind of problem reports
that appear when a bug in a critical piece of the software exists.

Thanks,
Shawn



Re: [SOLVED] ReplicationHandler - SnapPull failed to download a file completely.

2013-10-31 Thread Shalom Ben-Zvii Kazaz
Shawn, Thank you again for your help.

The problem appeared to be with httpclient.
I turned on debug logging for all libraries and saw a message "Garbage in
response" coming from httpclient just before the failure.
this is a log snippet:



31 Oct 2013 18:10:40,360 [explicit-fetchindex-cmd] DEBUG
DefaultClientConnection - Sending request: GET
/solr-master/replication?comman
d=filecontent&generation=6814&qt=%2Freplication&file=_aa7_Lucene41_0.pos&checksum=true&wt=filestream
HTTP/1.1
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - >> "GET
/solr-master/replication?command=filecontent&generation=6814&qt
=%2Freplication&file=_aa7_Lucene41_0.pos&checksum=true&wt=filestream
HTTP/1.1[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - >>
"User-Agent: Solr[org.apache.solr.client.solrj.impl.HttpSolrServer]
1.0[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - >> "Host:
solr-master.saltdev.sealdoc.com:8081[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - >>
"Connection: Keep-Alive[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - >>
"[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG headers - >> GET
/solr-master/replication?command=filecontent&generation=6814&
qt=%2Freplication&file=_aa7_Lucene41_0.pos&checksum=true&wt=filestream
HTTP/1.1
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG headers - >>
User-Agent: Solr[org.apache.solr.client.solrj.impl.HttpSolrServer
] 1.0
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG headers - >> Host:
solr-master.saltdev.sealdoc.com:8081
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG headers - >>
Connection: Keep-Alive
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - <<
"[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG
DefaultHttpResponseParser - Garbage in response:
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - <<
"4[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG
DefaultHttpResponseParser - Garbage in response: 4
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - <<
"[0x0][0x0][0x0][0x0][\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG
DefaultHttpResponseParser - Garbage in response: ^@^@^@^@
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - <<
"0[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG
DefaultHttpResponseParser - Garbage in response: 0
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG wire - <<
"[\r][\n]"
31 Oct 2013 18:10:40,361 [explicit-fetchindex-cmd] DEBUG
DefaultHttpResponseParser - Garbage in response:
31 Oct 2013 18:10:40,398 [explicit-fetchindex-cmd] DEBUG
DefaultClientConnection - Connection 0.0.0.0:55266<->172.16.77.121:8081closed
31 Oct 2013 18:10:40,398 [explicit-fetchindex-cmd] DEBUG
DefaultClientConnection - Connection 0.0.0.0:55266<->172.16.77.121:8081shut down
31 Oct 2013 18:10:40,398 [explicit-fetchindex-cmd] DEBUG
DefaultClientConnection - Connection 0.0.0.0:55266<->172.16.77.121:8081closed
31 Oct 2013 18:10:40,398 [explicit-fetchindex-cmd] DEBUG
PoolingClientConnectionManager - Connection released: [id: 0][route:
{}->http://solr-master.saltdev.sealdoc.com:8081][total kept alive: 1; route
allocated: 1 of 1; total allocated: 1 of 1]
31 Oct 2013 18:10:40,425 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Releasing directory:
/opt/watchdox/solr-slave/data/index 2 false
31 Oct 2013 18:10:40,425 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Reusing cached directory:
CachedDir<>
31 Oct 2013 18:10:40,425 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Releasing directory:
/opt/watchdox/solr-slave/data 0 false
31 Oct 2013 18:10:40,425 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Reusing cached directory:
CachedDir<>
31 Oct 2013 18:10:40,427 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Releasing directory:
/opt/watchdox/solr-slave/data 0 false
31 Oct 2013 18:10:40,428 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Done with dir:
CachedDir<>
31 Oct 2013 18:10:40,428 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Releasing directory:
/opt/watchdox/solr-slave/data/index.20131031180837277 0 true
31 Oct 2013 18:10:40,428 [explicit-fetchindex-cmd] INFO
CachingDirectoryFactory - looking to close
/opt/watchdox/solr-slave/data/index.20131031180837277
[CachedDir<>]
31 Oct 2013 18:10:40,428 [explicit-fetchindex-cmd] INFO
CachingDirectoryFactory - Closing directory:
/opt/watchdox/solr-slave/data/index.20131031180837277
31 Oct 2013 18:10:40,428 [explicit-fetchindex-cmd] INFO
CachingDirectoryFactory - Removing directory before core close:
/opt/watchdox/solr-slave/data/index.20131031180837277
31 Oct 2013 18:10:40,878 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Removing from cache:
CachedDir<>
31 Oct 2013 18:10:40,878 [explicit-fetchindex-cmd] DEBUG
CachingDirectoryFactory - Releasing director

dropping noise words and maintaining the relevancy

2013-10-31 Thread Susheel Kumar
Hello,

We have a very particular requirement of dropping noise words (LP, LLP, LLC, 
Corp, Corporation, Inc, Incoporation, PA, Professional Association, Attorney at 
law, GP, General Partnership etc.) at the end of search key but maintaining the 
relevancy. For e.g.

If user search for "160 Associates LP", we want search to return in their below 
relevancy order. Basically if exact / similar match is present, it comes first 
followed by other results.

160 Associates LP
160 Associates
160 Associates LLC
160 Associates LLLP
160 Hilton Associates

If I handle this through "Stop words" then LP will get dropped from search key 
and then all results will come but exact match will be shown somewhere lower or 
deep.

Regards and appreciate your help.
Susheel


Re: dropping noise words and maintaining the relevancy

2013-10-31 Thread Kranti Parisa
One possible approach is you can populate the titles in a field (say
exactMatch) and point your search query to exactMatch:"160 Associates LP"
OR text:""160 Associates LP"
assuming that you have all the text populated into the field called "text"

you can also use field level boosting with the above query, example
exactMatch:"160 Associates LP"^10 OR text:""160 Associates LP"^5


Thanks,
Kranti K. Parisa
http://www.linkedin.com/in/krantiparisa



On Thu, Oct 31, 2013 at 4:00 PM, Susheel Kumar <
susheel.ku...@thedigitalgroup.net> wrote:

> Hello,
>
> We have a very particular requirement of dropping noise words (LP, LLP,
> LLC, Corp, Corporation, Inc, Incoporation, PA, Professional Association,
> Attorney at law, GP, General Partnership etc.) at the end of search key but
> maintaining the relevancy. For e.g.
>
> If user search for "160 Associates LP", we want search to return in their
> below relevancy order. Basically if exact / similar match is present, it
> comes first followed by other results.
>
> 160 Associates LP
> 160 Associates
> 160 Associates LLC
> 160 Associates LLLP
> 160 Hilton Associates
>
> If I handle this through "Stop words" then LP will get dropped from search
> key and then all results will come but exact match will be shown somewhere
> lower or deep.
>
> Regards and appreciate your help.
> Susheel
>


Lock question

2013-10-31 Thread la...@protulae.com
Happy Halloween all,

 

I have a use case in which I have a custom indexer updating a Solr index. There 
is supposed to be no index updating via Solr itself - Solr is just a query tool 
in this case.

 

However, Solr insists on creating an UpdateHandler on startup, which creates a 
SolrIndexWriter, which then creates a lock file, and which therefore prevents 
my custom indexer from acquiring a lock. Sure, I can turn off locking which 
makes it all work, but I was just thinking: wouldn’t it be nice if we could 
have a NullUpdateHandler? Does that make sense?

 

Or, possibly there is a way for my custom indexer to somehow get the write lock 
away from Solr, but I don’t know how.

 

Thanks,

 

Lajos

HdfsDirectory Implementation

2013-10-31 Thread Josh Clum
Hello,

I refactored out the HDFS directory implementation from Solr to use in my
own project and was surprised to see how it performed. I'm using the both
the HDFSDirectory class and the
HdfsDirectoryFactory class.

On my local machine when using the cache there was a significant speed up.
It was a small enough that each file making up lucene index (12 docs) fit
into one block inside the cache.

When running it on a multinode cluster on AWS the performance pulling back
1031 docs with the cache was not that much better than without. According
to my log statements, the cache was being hit every time, but the
difference between this an my local was that there were several blocks per
file.

When setting up the cache I used the default setting as specified in
HdfsDirectoryFactory.

Any ideas on how to speed up searches? Should I change the block size? Is
there something that blur does to put a wrapper around the cache?

ON A MULTI NODE CLUSTER
Number of documents in directory[1031]
Try #1 -> Total execution time: 3776
Try #2 -> Total execution time: 2995
Try #3 -> Total execution time: 2683
Try #4 -> Total execution time: 2301
Try #5 -> Total execution time: 2174
Try #6 -> Total execution time: 2253
Try #7 -> Total execution time: 2184
Try #8 -> Total execution time: 2087
Try #9 -> Total execution time: 2157
Try #10 -> Total execution time: 2089
Cached try #1 -> Total execution time: 2065
Cached try #2 -> Total execution time: 2298
Cached try #3 -> Total execution time: 2398
Cached try #4 -> Total execution time: 2421
Cached try #5 -> Total execution time: 2080
Cached try #6 -> Total execution time: 2060
Cached try #7 -> Total execution time: 2285
Cached try #8 -> Total execution time: 2048
Cached try #9 -> Total execution time: 2087
Cached try #10 -> Total execution time: 2106

ON MY LOCAL
Number of documents in directory[12]
Try #1 -> Total execution time: 627
Try #2 -> Total execution time: 620
Try #3 -> Total execution time: 637
Try #4 -> Total execution time: 535
Try #5 -> Total execution time: 486
Try #6 -> Total execution time: 527
Try #7 -> Total execution time: 363
Try #8 -> Total execution time: 430
Try #9 -> Total execution time: 431
Try #10 -> Total execution time: 337
Cached try #1 -> Total execution time: 38
Cached try #2 -> Total execution time: 38
Cached try #3 -> Total execution time: 36
Cached try #4 -> Total execution time: 35
Cached try #5 -> Total execution time: 135
Cached try #6 -> Total execution time: 31
Cached try #7 -> Total execution time: 36
Cached try #8 -> Total execution time: 30
Cached try #9 -> Total execution time: 29
Cached try #10 -> Total execution time: 28

Thanks,
Josh


Re: Replacing Google Mini Search Appliance with Solr?

2013-10-31 Thread Furkan KAMACI
We also use Nutch at our environment.  Nutch crawls the data and it to Solr
for indexing. I have implemented a custom search API that interacts with my
Solr indexes cos of I don't want to expose my indexes directly to outside.
You can easily configure and build up what you want with such kind of
combination.

30 Ekim 2013 Çarşamba tarihinde Palmer, Eric  adlı
kullanıcı şöyle yazdı:
> Thanks for the link
>
> Sent from my iPhone
>
> On Oct 30, 2013, at 4:06 PM, "Rajani Maski"  wrote:
>
>> Hi Eric,
>>
>>  I have also developed mini-applications replacing GSA for some of our
>> clients using Apache Nutch + Solr to crawl multi lingual sites and enable
>> multi-lingual search. Nutch+Solr is very stable and Nutch mailing list
>> provides a good support.
>>
>> Reference link to start:
>> apache nutch | profilerajanimaski
>>
>> Thanks
>> Rajani
>>
>>
>>
>>
>> On Thu, Oct 31, 2013 at 12:27 AM, Palmer, Eric 
wrote:
>>
>>> Markus and Jason
>>>
>>> thanks for the info.
>>>
>>> I will start to research Nutch.  Writing a crawler, agree it is a rabbit
>>> hole.
>>>
>>>
>>> --
>>> Eric Palmer
>>>
>>> Web Services
>>> U of Richmond
>>>
>>> To report technical issues, obtain technical support or make requests
for
>>> enhancements please visit
>>> http://web.richmond.edu/contact/technical-support.html
>>>
>>>
>>>
>>>
>>>
>>> On 10/30/13 2:53 PM, "Jason Hellman" 
>>> wrote:
>>>
 Nutch is an excellent option.  It should feel very comfortable for
people
 migrating away from the Google appliances.

 Apache Droids is another possible way to approach, and I¹ve found
people
 using Heretrix or Manifold for various use cases (and usually in
 combination with other use cases where the extra overhead was worth the
 trouble).

 I think the simples approach will be NutchŠit¹s absolutely worth
taking a
 shot at it.

 DO NOT write a crawler!  That is a rabbit hole you do not want to peer
 down into :)



 On Oct 30, 2013, at 10:54 AM, Markus Jelsma 
 wrote:

> Hi Eric,
>
> We have also helped some government institution to replave their
> expensive GSA with open source software. In our case we use Apache
Nutch
> 1.7 to crawl the websites and index to Apache Solr. It is very
> effective, robust and scales easily with Hadoop if you have to. Nutch
> may not be the easiest tool for the job but is very stable, feature
rich
> and has an active community here at Apache.
>
> Cheers,
>
> -Original message-
>> From:Palmer, Eric 
>> Sent: Wednesday 30th October 2013 18:48
>> To: solr-user@lucene.apache.org
>> Subject: Replacing Google Mini Search Appliance with Solr?
>>
>> Hello all,
>>
>> Been lurking on the list for awhile.
>>
>> We are at the end of life for replacing two google mini search
>> appliances used to index our public web sites. Google is no longer
>> selling the mini appliances and buying the big appliance is not cost
>> beneficial.
>>
>> http://search.richmond.edu/
>>
>> We would run a solr replacement in linux (cents, redhat, similar)
with
>> open Java or Oracle Java.
>>
>> Background
>> ==
>> ~130 sites
>


Re: How to get similarity score between 0 and 1 not relative score

2013-10-31 Thread Toby Lazar
I think you are looking for something like this, though you can omit the fq
section:


http://localhost:8983/solr/collection/select?abc=text:bob&q={!func}scale(product(query($abc),1),0,1)&fq={!
frange l=0.9}$q

Also, I don't understand all the fuss about normalized scores.  In the
linked example, I can see an interest in searching for "apple bannana",
"zzz yyy xxx qqq kkk ttt rrr 111", etc. and wanting only close matches for
that point in time.  Would this be a good use for this approach?  I
understand that the results can change if the documents in the index change.

Thanks,

Toby



On Thu, Oct 31, 2013 at 12:56 AM, Anshum Gupta wrote:

> Hi Susheel,
>
> Have a look at this:
> http://wiki.apache.org/lucene-java/ScoresAsPercentages
>
> You may really want to reconsider doing that.
>
>
>
>
> On Thu, Oct 31, 2013 at 9:41 AM, sushil sharma  >wrote:
>
> > Hi,
> >
> > We have a requirement where user would like to see a score (between 0 to
> > 1) which can tell how close the input search string is with result
> string.
> > So if input was very close but not exact matach, score could be .90 etc.
> >
> > I do understand that we can get score from solr & divide by highest score
> > but that will always show 1 even if we match was not exact.
> >
> > Regards,
> > Susheel
>
>
>
>
> --
>
> Anshum Gupta
> http://www.anshumgupta.net
>


Simple (?) zookeeper question

2013-10-31 Thread Jack Park
Latest zookeeper is installed on an Ubuntu server box.
Java is 1.7 latest build.
whereis points to java just fine.
/etc/zookeeper is empty.

boot zookeeper from /bin as sudo ./zkServer.sh start
Console says "Started"
/etc/zookeeper now has a .pid file
In another console, ./zkServer.sh status returns:
"It's probably not running"

An interesting fact: the log4j.properties file says there should be a
zookeeper.log file in "."; there is no log file. When I do a text
search in the zookeeper source code for where it picks up the
log4j.properties, nothing is found.

Fascinating, what?  This must be a common beginner's question, not
well covered in web-search for my context. Does it ring any bells?

Many thanks.
Jack


Re: Error with CollapsingQParserPlugin when trying to use tagging

2013-10-31 Thread dboychuck
Here is an example URL that gives the error:

solr/productindex/productQuery?fq={!collapse%20field=groupid}&fq=discontinued:false&fq={!tag=manufacturer_string}manufacturer_string:(%22delta%22%20OR%20%22kohler%22)&fq=siteid:82&sort=score%20desc&facet=true&start=0&rows=48&fl=productid,manufacturer,uniqueFinish,uniqueid,productCompositeid,score&facet.query={!ex=onSale}onSale:true&facet.query={!ex=rating}rating:[4%20TO%20*]&facet.query={!ex=rating}rating:[3%20TO%20*]&facet.query={!ex=rating}rating:[2%20TO%20*]&facet.query={!ex=rating}rating:[1%20TO%20*]&facet.query={!ex=MadeinAmerica_boolean}MadeinAmerica_boolean:yes&facet.query={!ex=inStock}inStock:true&facet.query={!ex=PulloutSpray_string}PulloutSpray_string:yes&facet.query={!ex=HandlesIncluded_string}HandlesIncluded_string:yes&facet.query={!ex=Electronic_string}Electronic_string:yes&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[0%20TO%201]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[1%20TO%202]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[2%20TO%203]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[4%20TO%205]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[3%20TO%204]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[5%20TO%20*]&facet.query={!ex=ADA_string}ADA_string:yes&facet.query={!ex=WaterSenseCertified_string}WaterSenseCertified_string:yes&facet.query={!ex=WaterfallFaucet_boolean}WaterfallFaucet_boolean:yes&facet.query={!ex=InstallationAvailable_string}InstallationAvailable_string:yes&facet.query={!ex=LowLeadCompliant_string}LowLeadCompliant_string:yes&facet.query={!ex=DrainAssemblyIncluded_string}DrainAssemblyIncluded_string:yes&facet.query={!ex=EscutcheonIncluded_string}EscutcheonIncluded_string:yes&facet.field=NumberOfHandles_numeric&facet.field=pricebook_1_fs&facet.field=SpoutReach_numeric&facet.field=SpoutHeight_numeric&facet.field=FaucetCenters_numeric&facet.field=OverallHeight_numeric&facet.field=FaucetHoles_numeric&facet.field=HandleStyle_string&facet.field=masterFinish_string&facet.field={!ex=manufacturer_string}manufacturer_string&facet.field=HandleMaterial_string&facet.field=ValveType_string&facet.field=Theme_string&facet.field=MountingType_string&qt=/productQuery&qf=sku^9.0%20upc^9.1%20keywords_82_txtws^1.9%20uniqueid^9.0%20series^2.8%20productTitle^1.2%20productid^9.0%20manufacturer^4.0%20masterFinish^1.5%20theme^1.1%20categoryNames_82_txt^0.2%20finish^1.4&pf=keywords_82_txtws^2.1%20productTitle^1.5%20manufacturer^4.0%20finish^1.9&bf=linear(popularity_82_i,1,2)^3.0&q.alt=categories_82_is:108503

Taking out tagging I do not receive the error:

/solr/productindex/productQuery?fq={!collapse%20field=groupid}&fq=discontinued:false&fq=manufacturer_string:(%22delta%22%20OR%20%22kohler%22)&fq=siteid:82&sort=score%20desc&facet=true&start=0&rows=48&fl=productid,manufacturer,uniqueFinish,uniqueid,productCompositeid,score&facet.query={!ex=onSale}onSale:true&facet.query={!ex=rating}rating:[4%20TO%20*]&facet.query={!ex=rating}rating:[3%20TO%20*]&facet.query={!ex=rating}rating:[2%20TO%20*]&facet.query={!ex=rating}rating:[1%20TO%20*]&facet.query={!ex=MadeinAmerica_boolean}MadeinAmerica_boolean:yes&facet.query={!ex=inStock}inStock:true&facet.query={!ex=PulloutSpray_string}PulloutSpray_string:yes&facet.query={!ex=HandlesIncluded_string}HandlesIncluded_string:yes&facet.query={!ex=Electronic_string}Electronic_string:yes&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[0%20TO%201]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[1%20TO%202]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[2%20TO%203]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[4%20TO%205]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[3%20TO%204]&facet.query={!ex=FlowRateGPM_numeric}FlowRateGPM_numeric:[5%20TO%20*]&facet.query={!ex=ADA_string}ADA_string:yes&facet.query={!ex=WaterSenseCertified_string}WaterSenseCertified_string:yes&facet.query={!ex=WaterfallFaucet_boolean}WaterfallFaucet_boolean:yes&facet.query={!ex=InstallationAvailable_string}InstallationAvailable_string:yes&facet.query={!ex=LowLeadCompliant_string}LowLeadCompliant_string:yes&facet.query={!ex=DrainAssemblyIncluded_string}DrainAssemblyIncluded_string:yes&facet.query={!ex=EscutcheonIncluded_string}EscutcheonIncluded_string:yes&facet.field=NumberOfHandles_numeric&facet.field=pricebook_1_fs&facet.field=SpoutReach_numeric&facet.field=SpoutHeight_numeric&facet.field=FaucetCenters_numeric&facet.field=OverallHeight_numeric&facet.field=FaucetHoles_numeric&facet.field=HandleStyle_string&facet.field=masterFinish_string&facet.field={!ex=manufacturer_string}manufacturer_string&facet.field=HandleMaterial_string&facet.field=ValveType_string&facet.field=Theme_string&facet.field=MountingType_string&qt=/productQuery&qf=sku^9.0%20upc^9.1%20keywords_82_txtws^1.9%20uniqueid^9.0%20series^2.8%20productTitle^1.2%20productid^9.0%20manufacturer^4.0%20masterFinish^1.5%20theme^1.1%20categoryNames_82_txt^0.2%20finish^1.4&pf=keywords_82_t

Re: How to get similarity score between 0 and 1 not relative score

2013-10-31 Thread sushil sharma
Thanks, Anshum. That really helps. Now my question is
 
Our mid to long term vision is to automate the searches and send results to 
users which have higher score. For e.g. records with score higher than 90% or 
so, we would like to include in a report and send across to users. But since 
the score is relative or it can be 1.5 or 6.5 or any number we don't know if 
the result we are sending is really relevant or a close match or not. 
 
How can we deal with this situation.
 
Regards & Appreciate your help.
Susheel



On Wednesday, 30 October 2013 11:57 PM, Anshum Gupta  
wrote:
  
Hi Susheel,

Have a look at this: http://wiki.apache.org/lucene-java/ScoresAsPercentages

You may really want to reconsider doing that.  





On Thu, Oct 31, 2013 at 9:41 AM, sushil sharma  wrote:

Hi,
> 
>We have a requirement where user would like to see a score (between 0 to 1) 
>which can tell how close the input search string is with result string. So if 
>input was very close but not exact matach, score could be .90 etc.
> 
>I do understand that we can get score from solr & divide by highest score but 
>that will always show 1 even if we match was not exact.
> 
>Regards,
>Susheel


-- 

Anshum Gupta
http://www.anshumgupta.net/ 

Error with CollapsingQParserPlugin when trying to use tagging

2013-10-31 Thread dboychuck
I ran into an error with the CollapsingQParserPlugin when trying to use it in
tandem with tagging

I get the following error whenever I use {!tag} in the same request as
{!collapse field=groupid}

Oct 31, 2013 6:43:56 PM org.apache.tomcat.util.http.Cookies
processCookieHeader
INFO: Cookies: Invalid cookie. Value not a token or quoted value
 Note: further occurrences of Cookie errors will be logged at DEBUG level.
121624 [http-bio-8080-exec-1] ERROR org.apache.solr.core.SolrCore  –
java.lang.NullPointerException
at
org.apache.solr.search.CollapsingQParserPlugin$CollapsingScoreCollector.collect(CollapsingQParserPlugin.java:409)
at
org.apache.solr.search.SolrIndexSearcher.getDocSet(SolrIndexSearcher.java:910)
at 
org.apache.solr.request.SimpleFacets.parseParams(SimpleFacets.java:219)
at
org.apache.solr.request.SimpleFacets.getFacetFieldCounts(SimpleFacets.java:549)
at
org.apache.solr.request.SimpleFacets.getFacetCounts(SimpleFacets.java:265)
at
org.apache.solr.handler.component.FacetComponent.process(FacetComponent.java:78)
at
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:208)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1859)
at
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:703)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:406)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:195)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:243)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:210)
at
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:222)
at
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:123)
at
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:171)
at
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:99)
at
org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:953)
at
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118)
at
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:408)
at
org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1008)
at
org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:589)
at
org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:310)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
at java.lang.Thread.run(Thread.java:722)

121635 [http-bio-8080-exec-1] ERROR
org.apache.solr.servlet.SolrDispatchFilter  –
null:java.lang.NullPointerException
at
org.apache.solr.search.CollapsingQParserPlugin$CollapsingScoreCollector.collect(CollapsingQParserPlugin.java:409)
at
org.apache.solr.search.SolrIndexSearcher.getDocSet(SolrIndexSearcher.java:910)
at 
org.apache.solr.request.SimpleFacets.parseParams(SimpleFacets.java:219)
at
org.apache.solr.request.SimpleFacets.getFacetFieldCounts(SimpleFacets.java:549)
at
org.apache.solr.request.SimpleFacets.getFacetCounts(SimpleFacets.java:265)
at
org.apache.solr.handler.component.FacetComponent.process(FacetComponent.java:78)
at
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:208)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1859)
at
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:703)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:406)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:195)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:243)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:210)
at
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:222)
at
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:123)
at
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:171)
at
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:99)
at
org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:953)
at
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118)
at
or

RE: dropping noise words and maintaining the relevancy

2013-10-31 Thread Susheel Kumar
Thanks, Kranti. Nice suggestion. I'll try it out. 

-Original Message-
From: Kranti Parisa [mailto:kranti.par...@gmail.com] 
Sent: Thursday, October 31, 2013 3:18 PM
To: solr-user@lucene.apache.org
Subject: Re: dropping noise words and maintaining the relevancy

One possible approach is you can populate the titles in a field (say
exactMatch) and point your search query to exactMatch:"160 Associates LP"
OR text:""160 Associates LP"
assuming that you have all the text populated into the field called "text"

you can also use field level boosting with the above query, example
exactMatch:"160 Associates LP"^10 OR text:""160 Associates LP"^5


Thanks,
Kranti K. Parisa
http://www.linkedin.com/in/krantiparisa



On Thu, Oct 31, 2013 at 4:00 PM, Susheel Kumar < 
susheel.ku...@thedigitalgroup.net> wrote:

> Hello,
>
> We have a very particular requirement of dropping noise words (LP, 
> LLP, LLC, Corp, Corporation, Inc, Incoporation, PA, Professional 
> Association, Attorney at law, GP, General Partnership etc.) at the end 
> of search key but maintaining the relevancy. For e.g.
>
> If user search for "160 Associates LP", we want search to return in 
> their below relevancy order. Basically if exact / similar match is 
> present, it comes first followed by other results.
>
> 160 Associates LP
> 160 Associates
> 160 Associates LLC
> 160 Associates LLLP
> 160 Hilton Associates
>
> If I handle this through "Stop words" then LP will get dropped from 
> search key and then all results will come but exact match will be 
> shown somewhere lower or deep.
>
> Regards and appreciate your help.
> Susheel
>


unable to load core after cluster restart

2013-10-31 Thread kaustubh147
Hi, 

Glassfish 3.1.2.2 
Solr 4.5 
Zookeeper 3.4.5 

We have set up a SolrCloud with 4 Solr nodes and 3 zookeeper instances. 

I start the cluster for the first time with bootstrap_conf= true All the
nodes starts property.. I am creating cores (with the same name) on all 4
instances. I can add multiple cores on each of the instances... logically I
have 5 collections.

Now i am creating indexes.. and it automatically creates 4 copies of the
index, one for each instance in appropriate SolrHome directory... It will
work properly untill I restart the Solr cluster

as soon as I restart the cluster, it throws this error (refer below) and
none of the collection works properly...


ERROR - 2013-10-31 19:23:24.411; org.apache.solr.core.CoreContainer; Unable
to create core: xyz
org.apache.solr.common.SolrException: Error opening new searcher
at org.apache.solr.core.SolrCore.(SolrCore.java:834)
at org.apache.solr.core.SolrCore.(SolrCore.java:625)
at org.apache.solr.core.ZkContainer.createFromZk(ZkContainer.java:256)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:557)
at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:249)
at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:241)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)
Caused by: org.apache.solr.common.SolrException: Error opening new searcher
at org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:1477)
at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1589)
at org.apache.solr.core.SolrCore.(SolrCore.java:821)
... 13 more
Caused by: org.apache.lucene.store.LockObtainFailedException: Lock obtain
timed out:
NativeFSLock@/mnt/emc/app_name/data-refresh/SolrCloud/SolrHome1/solr/xyz/data/index/write.lock
at org.apache.lucene.store.Lock.obtain(Lock.java:84)
at org.apache.lucene.index.IndexWriter.(IndexWriter.java:673)
at 
org.apache.solr.update.SolrIndexWriter.(SolrIndexWriter.java:77)
at 
org.apache.solr.update.SolrIndexWriter.create(SolrIndexWriter.java:64)
at
org.apache.solr.update.DefaultSolrCoreState.createMainIndexWriter(DefaultSolrCoreState.java:267)
at
org.apache.solr.update.DefaultSolrCoreState.getIndexWriter(DefaultSolrCoreState.java:110)
at org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:1440)
... 15 more
ERROR - 2013-10-31 19:23:24.420; org.apache.solr.common.SolrException;
null:org.apache.solr.common.SolrException: Unable to create core: xyz
at
org.apache.solr.core.CoreContainer.recordAndThrow(CoreContainer.java:936)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:568)
at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:249)
at org.apache.solr.core.CoreContainer$1.call(CoreContainer.java:241)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)
Caused by: org.apache.solr.common.SolrException: Error opening new searcher
at org.apache.solr.core.SolrCore.(SolrCore.java:834)
at org.apache.solr.core.SolrCore.(SolrCore.java:625)
at org.apache.solr.core.ZkContainer.createFromZk(ZkContainer.java:256)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:557)
... 10 more
Caused by: org.apache.solr.common.SolrException: Error opening new searcher
at org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:1477)
at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1589)
at org.apache.solr.core.SolrCore.(SolrCore.java:821)
... 13 more
Caused by: org.apache.lucene.store.LockObtainFailedException: Lock obtain
timed out:
NativeFSLock@/mnt/emc/app_name/data-refresh/SolrCloud/SolrHome1/solr/xyz/data/index/write.lock
at org.apache.lucene.store.Lock.obtain(Lock.java:84)
at org.apache.lucene.index

Re: Problem with glassfish and zookeeper 3.4.5

2013-10-31 Thread kaustubh147
Thanks Shawn,

I found a bug in my code, it was creating too many CloudSolrServer objects.

Thanks
Kaustubh



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Problem-with-glassfish-and-zookeeper-3-4-5-tp4097635p4098732.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Simple (?) zookeeper question

2013-10-31 Thread Jack Park
After digging deeper (slow for a *nix newbee), I uncovered issues with
the java installation. A step in installation of Oracle Java has it
that you -install "java" with the path to /bin/java. That done,
zookeeper seems to be running.

I booted three cores (on the same box) -- this is the simple one-box
3-node cloud test, and used the test code from the Lucidworks course
to send over and read some documents. That failed with this:
Unknown document router '{name=compositeId}'

Lots more research.
Closer...

On Thu, Oct 31, 2013 at 5:44 PM, Jack Park  wrote:
> Latest zookeeper is installed on an Ubuntu server box.
> Java is 1.7 latest build.
> whereis points to java just fine.
> /etc/zookeeper is empty.
>
> boot zookeeper from /bin as sudo ./zkServer.sh start
> Console says "Started"
> /etc/zookeeper now has a .pid file
> In another console, ./zkServer.sh status returns:
> "It's probably not running"
>
> An interesting fact: the log4j.properties file says there should be a
> zookeeper.log file in "."; there is no log file. When I do a text
> search in the zookeeper source code for where it picks up the
> log4j.properties, nothing is found.
>
> Fascinating, what?  This must be a common beginner's question, not
> well covered in web-search for my context. Does it ring any bells?
>
> Many thanks.
> Jack


Re: unable to load core after cluster restart

2013-10-31 Thread Shawn Heisey
On 10/31/2013 9:18 PM, kaustubh147 wrote:
> Glassfish 3.1.2.2 
> Solr 4.5 
> Zookeeper 3.4.5 
> 
> We have set up a SolrCloud with 4 Solr nodes and 3 zookeeper instances. 
> 
> I start the cluster for the first time with bootstrap_conf= true All the
> nodes starts property.. I am creating cores (with the same name) on all 4
> instances. I can add multiple cores on each of the instances... logically I
> have 5 collections.
> 
> Now i am creating indexes.. and it automatically creates 4 copies of the
> index, one for each instance in appropriate SolrHome directory... It will
> work properly untill I restart the Solr cluster
> 
> as soon as I restart the cluster, it throws this error (refer below) and
> none of the collection works properly...

It's having problems with the index locking.  Further down, you talk
about shared directories.  I assume you're using a network filesystem,
like NFS.  Lucene and Solr don't work well with network file systems,
and NFS in particular.

http://stackoverflow.com/questions/9599529/solr-over-nfs-problems

For SolrCloud, using the bootstrap options tends to create a lot of
confusion.

> 1. I have removed dataDir from solrconfig.xml as suggested by Shaun here...
> 
> http://lucene.472066.n3.nabble.com/Solr-4-3-0-Shard-instances-using-incorrect-data-directory-on-machine-boot-td4063799.html
> 
> 2. I have provided absolute dataDir path in the core.properties file -
> https://issues.apache.org/jira/browse/SOLR-4878
> 
> 3. InstanceDir in each SolrHome have same name for every core/collection--
> for example
> 
> SolrHome1/solr/xyz/conf
> SolrHome1/solr/xyz/data
> SolrHome1/solr/xyz/core.properties
> SolrHome1/solr/pqr/conf
> SolrHome1/solr/pqr/data
> SolrHome1/solr/pqr/core.properties
> 
> 
> SolrHome2/solr/xyz/conf
> SolrHome2/solr/xyz/data
> SolrHome2/solr/xyz/core.properties
> SolrHome2/solr/pqr/conf
> SolrHome2/solr/pqr/data
> SolrHome2/solr/pqr/core.properties
> 
> ...
> 
> 3. The 4 SolrHome for each of the instances are on a single shared drive...
> but are in different directories
> 
> 4. All my collections and cores share the same solrconfig.xml 

The fact that you're using SolrCloud changes things quite a bit.  I
notice that you have a "conf" directory on all your cores.  I'm very
curious as to why -- because with SolrCloud, the config that's used
isn't on the disk, it's in zookeeper ... and when you create a
collection from scratch, the cores will NOT contain a conf directory.

IMHO, you shouldn't be trying to control things like dataDir with
SolrCloud.  In fact, you should let SolrCloud control practically all
aspects of your cores.  Here's some more stuff that's my opinion about
how a SolrCloud should be begun and managed:

Set the solr home, either with a solr.solr.home java system property or
the JNDI property solr/home.  In solr.xml, include settings for the
hostPort and hostContext, and include a zkHost parameter as well with
the following format, where "mysolr" is a zookeeper chroot and can be
anything you want it to be:

zoo1.example.com:2181,zoo2.example.com:2181,zoo3.example.com/mysolr

Upload one or more configs to zookeeper using the zkcli "upconfig" command.

The first time you start each Solr server, it should have no cores at
all, and you should never use any bootstrap options.

Once all servers are started, create a collection using the following
URL as a template:

http://server:port/admin/collections?action=CREATE&name=mytest&numShards=X&replicationFactor=N&collection.configName=mycfg

Use appropriate values for name, X and N, and N should be at least 2
unless you're building a dev cluster that doesn't need redundancy.  For
the collection.configName, use a config name that you uploaded earlier
with zkcli.

Thanks,
Shawn



Re: character encoding issue...

2013-10-31 Thread Chris
Hi Rajani,

I followed the steps exactly as in
http://zensarteam.wordpress.com/2011/11/25/6-steps-to-configure-solr-on-apache-tomcat-7-0-20/

However, when i send a query to this new instance in tomcat, i again get
the error -

  Scheduled Groups Maintenance
In preparation for the new release roll-out, Diigo groups won’t be
accessible on Sept 28 (Mon) around midnight 0:00 PST for several
hours.
Stay tuned to say hello to Diigo V4 soon!

location of the text  -
http://blog.diigo.com/2009/09/28/scheduled-groups-maintenance/

same problem at - http://cn.nytimes.com/business/20130926/c26alibaba/

All text in title comes like -

 - � 

   -
� 



Can you please advice?

Chris




On Tue, Oct 29, 2013 at 11:33 PM, Rajani Maski wrote:

> Hi,
>
>If you are using Apache Tomcat Server, hope you are not missing the
> below mentioned configuration:
>
>   connectionTimeout=”2″
> redirectPort=”8443″ *URIEncoding=”UTF-8″*/>
>
> I had faced similar issue with Chinese Characters and had resolved with the
> above config.
>
> Links for reference :
>
> http://zensarteam.wordpress.com/2011/11/25/6-steps-to-configure-solr-on-apache-tomcat-7-0-20/
>
> http://blog.sidu.in/2007/05/tomcat-and-utf-8-encoded-uri-parameters.html#.Um_3P3Cw2X8
>
>
> Thanks
>
>
>
> On Tue, Oct 29, 2013 at 9:20 PM, Chris  wrote:
>
> > Hi All,
> >
> > I get characters like -
> >
> > �� - CTA -
> >
> > in the solr index. I am adding Java beans to solr by the addBean()
> > function.
> >
> > This seems to be a character encoding issue. Any pointers on how to
> > resolve this one?
> >
> > I have seen that this occurs  mostly for japanese chinese characters.
> >
>