Hmm...this is very strange - there is nothing interesting in any of the logs?
In clusterstate.json, all of the shards have an active state? There are quite a few of us doing exactly this setup recently, so there must be something we are missing here... Any info you can offer might help. - Mark On Feb 28, 2012, at 1:00 PM, Matthew Parker wrote: > Mark, > > I got the codebase from the 2/26/2012, and I got the same inconsistent > results. > > I have solr running on four ports 8081-8084 > > 8081 and 8082 are the leaders for shard 1, and shard 2, respectively > > 8083 - is assigned to shard 1 > 8084 - is assigned to shard 2 > > queries come in and sometime it seems the windows from 8081 and 8083 move > responding to the query but there are no results. > > if the queries run on 8081/8082 or 8081/8084 then results come back ok. > > The query is nothing more than: q=*:* > > Regards, > > Matt > > > On Mon, Feb 27, 2012 at 9:26 PM, Matthew Parker < > mpar...@apogeeintegration.com> wrote: > >> I'll have to check on the commit situation. We have been pushing data from >> SharePoint the last week or so. Would that somehow block the documents >> moving between the solr instances? >> >> I'll try another version tomorrow. Thanks for the suggestions. >> >> On Mon, Feb 27, 2012 at 5:34 PM, Mark Miller <markrmil...@gmail.com>wrote: >> >>> Hmmm...all of that looks pretty normal... >>> >>> Did a commit somehow fail on the other machine? When you view the stats >>> for the update handler, are there a lot of pending adds for on of the >>> nodes? Do the commit counts match across nodes? >>> >>> You can also query an individual node with distrib=false to check that. >>> >>> If you build is a month old, I'd honestly recommend you try upgrading as >>> well. >>> >>> - Mark >>> >>> On Feb 27, 2012, at 3:34 PM, Matthew Parker wrote: >>> >>>> Here is most of the cluster state: >>>> >>>> Connected to Zookeeper >>>> localhost:2181, localhost: 2182, localhost:2183 >>>> >>>> /(v=0 children=7) "" >>>> /CONFIGS(v=0, children=1) >>>> /CONFIGURATION(v=0 children=25) >>>> <<<<< all the configuration files, velocity info, xslt, etc. >>>>>>>> >>>> /NODE_STATES(v=0 children=4) >>>> MACHINE1:8083_SOLR (v=121)"[{"shard_id":"shard1", >>>> "state":"active","core":"","collection":"collection1","node_name:"..." >>>> MACHINE1:8082_SOLR (v=101)"[{"shard_id":"shard2", >>>> "state":"active","core":"","collection":"collection1","node_name:"..." >>>> MACHINE1:8081_SOLR (v=92)"[{"shard_id":"shard1", >>>> "state":"active","core":"","collection":"collection1","node_name:"..." >>>> MACHINE1:8084_SOLR (v=73)"[{"shard_id":"shard2", >>>> "state":"active","core":"","collection":"collection1","node_name:"..." >>>> /ZOOKEEPER (v-0 children=1) >>>> QUOTA(v=0) >>>> >>>> >>> /CLUSTERSTATE.JSON(V=272)"{"collection1":{"shard1":{MACHINE1:8081_solr_":{shard_id":"shard1","leader":"true","..." >>>> /LIVE_NODES (v=0 children=4) >>>> MACHINE1:8083_SOLR(ephemeral v=0) >>>> MACHINE1:8082_SOLR(ephemeral v=0) >>>> MACHINE1:8081_SOLR(ephemeral v=0) >>>> MACHINE1:8084_SOLR(ephemeral v=0) >>>> /COLLECTIONS (v=1 children=1) >>>> COLLECTION1(v=0 children=2)"{"configName":"configuration1"}" >>>> LEADER_ELECT(v=0 children=2) >>>> SHARD1(V=0 children=1) >>>> ELECTION(v=0 children=2) >>>> >>>> 87186203314552835-MACHINE1:8081_SOLR_-N_0000000096(ephemeral v=0) >>>> >>>> 87186203314552836-MACHINE1:8083_SOLR_-N_0000000084(ephemeral v=0) >>>> SHARD2(v=0 children=1) >>>> ELECTION(v=0 children=2) >>>> >>>> 231301391392833539-MACHINE1:8084_SOLR_-N_0000000085(ephemeral v=0) >>>> >>>> 159243797356740611-MACHINE1:8082_SOLR_-N_0000000084(ephemeral v=0) >>>> LEADERS (v=0 children=2) >>>> SHARD1 (ephemeral >>>> v=0)"{"core":"","node_name":"MACHINE1:8081_solr","base_url":" >>>> http://MACHINE1:8081/solr"}" >>>> SHARD2 (ephemeral >>>> v=0)"{"core":"","node_name":"MACHINE1:8082_solr","base_url":" >>>> http://MACHINE1:8082/solr"}" >>>> /OVERSEER_ELECT (v=0 children=2) >>>> ELECTION (v=0 children=4) >>>> 231301391392833539-MACHINE1:8084_SOLR_-N_0000000251(ephemeral >>> v=0) >>>> 87186203314552835-MACHINE1:8081_SOLR_-N_0000000248(ephemeral >>> v=0) >>>> 159243797356740611-MACHINE1:8082_SOLR_-N_0000000250(ephemeral >>> v=0) >>>> 87186203314552836-MACHINE1:8083_SOLR_-N_0000000249(ephemeral >>> v=0) >>>> LEADER (emphemeral >>>> v=0)"{"id":"87186203314552835-MACHINE1:8081_solr-n_000000248"}" >>>> >>>> >>>> >>>> On Mon, Feb 27, 2012 at 2:47 PM, Mark Miller <markrmil...@gmail.com> >>> wrote: >>>> >>>>> >>>>> On Feb 27, 2012, at 2:22 PM, Matthew Parker wrote: >>>>> >>>>>> Thanks for your reply Mark. >>>>>> >>>>>> I believe the build was towards the begining of the month. The >>>>>> solr.spec.version is 4.0.0.2012.01.10.38.09 >>>>>> >>>>>> I cannot access the clusterstate.json contents. I clicked on it a >>> couple >>>>> of >>>>>> times, but nothing happens. Is that stored on disk somewhere? >>>>> >>>>> Are you using the new admin UI? That has recently been updated to work >>>>> better with cloud - it had some troubles not too long ago. If you are, >>> you >>>>> should trying using the old admin UI's zookeeper page - that should >>> show >>>>> the cluster state. >>>>> >>>>> That being said, there has been a lot of bug fixes over the past month >>> - >>>>> so you may just want to update to a recent version. >>>>> >>>>>> >>>>>> I configured a custom request handler to calculate an unique document >>> id >>>>>> based on the file's url. >>>>>> >>>>>> On Mon, Feb 27, 2012 at 1:13 PM, Mark Miller <markrmil...@gmail.com> >>>>> wrote: >>>>>> >>>>>>> Hey Matt - is your build recent? >>>>>>> >>>>>>> Can you visit the cloud/zookeeper page in the admin and send the >>>>> contents >>>>>>> of the clusterstate.json node? >>>>>>> >>>>>>> Are you using a custom index chain or anything out of the ordinary? >>>>>>> >>>>>>> >>>>>>> - Mark >>>>>>> >>>>>>> On Feb 27, 2012, at 12:26 PM, Matthew Parker wrote: >>>>>>> >>>>>>>> TWIMC: >>>>>>>> >>>>>>>> Environment >>>>>>>> ========= >>>>>>>> Apache SOLR rev-1236154 >>>>>>>> Apache Zookeeper 3.3.4 >>>>>>>> Windows 7 >>>>>>>> JDK 1.6.0_23.b05 >>>>>>>> >>>>>>>> I have built a SOLR Cloud instance with 4 nodes using the embeded >>> Jetty >>>>>>>> servers. >>>>>>>> >>>>>>>> I created a 3 node zookeeper ensemble to manage the solr >>> configuration >>>>>>> data. >>>>>>>> >>>>>>>> All the instances run on one server so I've had to move ports around >>>>> for >>>>>>>> the various applications. >>>>>>>> >>>>>>>> I start the 3 zookeeper nodes. >>>>>>>> >>>>>>>> I started the first instance of solr cloud with the parameter to >>> have >>>>> two >>>>>>>> shards. >>>>>>>> >>>>>>>> The start the remaining 3 solr nodes. >>>>>>>> >>>>>>>> The system comes up fine. No errors thrown. >>>>>>>> >>>>>>>> I can view the solr cloud console and I can see the SOLR >>> configuration >>>>>>>> files managed by ZooKeeper. >>>>>>>> >>>>>>>> I published data into the SOLR Cloud instances from SharePoint using >>>>>>> Apache >>>>>>>> Manifold 0.4-incubating. Manifold is setup to publish the data into >>>>>>>> collection1, which is the only collection defined in the cluster. >>>>>>>> >>>>>>>> When I query the data from collection1 as per the solr wiki, the >>>>> results >>>>>>>> are inconsistent. Sometimes all the results are there, other times >>>>>>> nothing >>>>>>>> comes back at all. >>>>>>>> >>>>>>>> It seems to be having an issue auto replicating the data across the >>>>>>> cloud. >>>>>>>> >>>>>>>> Is there some specific setting I might have missed? Based upon what >>> I >>>>>>> read, >>>>>>>> I thought that SOLR cloud would take care of distributing and >>>>> replicating >>>>>>>> the data automatically. Do you have to tell it what shard to publish >>>>> the >>>>>>>> data into as well? >>>>>>>> >>>>>>>> Any help would be appreciated. >>>>>>>> >>>>>>>> Thanks, >>>>>>>> >>>>>>>> Matt >>>>>>>> >>>>>>>> ------------------------------ >>>>>>>> This e-mail and any files transmitted with it may be proprietary. >>>>>>> Please note that any views or opinions presented in this e-mail are >>>>> solely >>>>>>> those of the author and do not necessarily represent those of Apogee >>>>>>> Integration. >>>>>>> >>>>>>> - Mark Miller >>>>>>> lucidimagination.com >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> Matt >>>>>> >>>>>> ------------------------------ >>>>>> This e-mail and any files transmitted with it may be proprietary. >>>>> Please note that any views or opinions presented in this e-mail are >>> solely >>>>> those of the author and do not necessarily represent those of Apogee >>>>> Integration. >>>>> >>>>> - Mark Miller >>>>> lucidimagination.com >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>> >>>> ------------------------------ >>>> This e-mail and any files transmitted with it may be proprietary. >>> Please note that any views or opinions presented in this e-mail are solely >>> those of the author and do not necessarily represent those of Apogee >>> Integration. >>> >>> - Mark Miller >>> lucidimagination.com >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >> > > ------------------------------ > This e-mail and any files transmitted with it may be proprietary. Please > note that any views or opinions presented in this e-mail are solely those of > the author and do not necessarily represent those of Apogee Integration. - Mark Miller lucidimagination.com