I tried publishing to /update/extract request handler using manifold, but got the same result.
I also tried swapping out the replication handlers too, but that didn't do anything. Otherwise, that's it. On Thu, Mar 1, 2012 at 3:35 PM, Mark Miller <markrmil...@gmail.com> wrote: > Any other customizations you are making to solrconfig? > > On Mar 1, 2012, at 1:48 PM, Matthew Parker wrote: > > > Added it back in. I still get the same result. > > > > On Wed, Feb 29, 2012 at 10:09 PM, Mark Miller <markrmil...@gmail.com> > wrote: > > Do you have a _version_ field in your schema? I actually just came back > to > > this thread with that thought and then saw your error - so that remains > my > > guess. > > > > I'm going to improve the doc on the wiki around what needs to be defined > > for SolrCloud - so far we have things in the example defaults, but its > not > > clear enough to users what needs to be there if they are using an old > > schema or modifying the example. > > > > <field name="_version_" type="long" indexed="true" stored="true"/> > > > > If you do have a _version_ field, there is something to track down here > for > > sure. > > > > On Wed, Feb 29, 2012 at 1:15 PM, Matthew Parker < > > mpar...@apogeeintegration.com> wrote: > > > > > Mark/Sami > > > > > > I ran the system with 3 zookeeper nodes, 2 solr cloud nodes, and left > > > numShards set to its default value (i.e. 1) > > > > > > I looks like it finally sync'd with the other one after quite a while, > but > > > it's throwing lots of errors like the following: > > > > > > org.apache.solr.common.SolrException: missing _version_ on update from > > > leader at > > > > org.apache.solr.update.processor.DistributtedUpdateProcessor.versionDelete( > > > DistributedUpdateProcessor.java: 712) > > > .... > > > .... > > > .... > > > > > > Is it normal to sync long after the documents were sent for indexing? > > > > > > I'll have to check and see whether the 4 solr node instance with 2 > shards > > > works after waiting for the system to sync. > > > > > > Regards, > > > > > > Matt > > > > > > On Wed, Feb 29, 2012 at 12:03 PM, Matthew Parker < > > > mpar...@apogeeintegration.com> wrote: > > > > > > > I also took out my requestHandler and used the standard > /update/extract > > > > handler. Same result. > > > > > > > > On Wed, Feb 29, 2012 at 11:47 AM, Matthew Parker < > > > > mpar...@apogeeintegration.com> wrote: > > > > > > > >> I tried running SOLR Cloud with the default number of shards (i.e. > 1), > > > >> and I get the same results. > > > >> > > > >> On Wed, Feb 29, 2012 at 10:46 AM, Matthew Parker < > > > >> mpar...@apogeeintegration.com> wrote: > > > >> > > > >>> Mark, > > > >>> > > > >>> Nothing appears to be wrong in the logs. I wiped the indexes and > > > >>> imported 37 files from SharePoint using Manifold. All 37 make it > in, > > > but > > > >>> SOLR still has issues with the results being inconsistent. > > > >>> > > > >>> Let me run my setup by you, and see whether that is the issue? > > > >>> > > > >>> On one machine, I have three zookeeper instances, four solr > instances, > > > >>> and a data directory for solr and zookeeper config data. > > > >>> > > > >>> Step 1. I modified each zoo.xml configuration file to have: > > > >>> > > > >>> Zookeeper 1 - Create /zookeeper1/conf/zoo.cfg > > > >>> ================ > > > >>> tickTime=2000 > > > >>> initLimit=10 > > > >>> syncLimit=5 > > > >>> dataDir=[DATA_DIRECTORY]/zk1_data > > > >>> clientPort=2181 > > > >>> server.1=localhost:2888:3888 > > > >>> server.2=localhost:2889:3889 > > > >>> server.3=localhost:2890:3890 > > > >>> > > > >>> Zookeeper 1 - Create /[DATA_DIRECTORY]/zk1_data/myid with the > following > > > >>> contents: > > > >>> ============================================================== > > > >>> 1 > > > >>> > > > >>> Zookeep 2 - Create /zookeeper2/conf/zoo.cfg > > > >>> ============== > > > >>> tickTime=2000 > > > >>> initLimit=10 > > > >>> syncLimit=5 > > > >>> dataDir=[DATA_DIRECTORY]/zk2_data > > > >>> clientPort=2182 > > > >>> server.1=localhost:2888:3888 > > > >>> server.2=localhost:2889:3889 > > > >>> server.3=localhost:2890:3890 > > > >>> > > > >>> Zookeeper 2 - Create /[DATA_DIRECTORY]/zk2_data/myid with the > following > > > >>> contents: > > > >>> ============================================================== > > > >>> 2 > > > >>> > > > >>> Zookeeper 3 - Create /zookeeper3/conf/zoo.cfg > > > >>> ================ > > > >>> tickTime=2000 > > > >>> initLimit=10 > > > >>> syncLimit=5 > > > >>> dataDir=[DATA_DIRECTORY]/zk3_data > > > >>> clientPort=2183 > > > >>> server.1=localhost:2888:3888 > > > >>> server.2=localhost:2889:3889 > > > >>> server.3=localhost:2890:3890 > > > >>> > > > >>> Zookeeper 3 - Create /[DATA_DIRECTORY]/zk3_data/myid with the > following > > > >>> contents: > > > >>> ==================================================== > > > >>> 3 > > > >>> > > > >>> Step 2 - SOLR Build > > > >>> =============== > > > >>> > > > >>> I pulled the latest SOLR trunk down. I built it with the following > > > >>> commands: > > > >>> > > > >>> ant example dist > > > >>> > > > >>> I modified the solr.war files and added the solr cell and > extraction > > > >>> libraries to WEB-INF/lib. I couldn't get the extraction to work > > > >>> any other way. Will zookeper pickup jar files stored with the rest > of > > > >>> the configuration files in Zookeeper? > > > >>> > > > >>> I copied the contents of the example directory to each of my SOLR > > > >>> directories. > > > >>> > > > >>> Step 3 - Starting Zookeeper instances > > > >>> =========================== > > > >>> > > > >>> I ran the following commands to start the zookeeper instances: > > > >>> > > > >>> start .\zookeeper1\bin\zkServer.cmd > > > >>> start .\zookeeper2\bin\zkServer.cmd > > > >>> start .\zookeeper3\bin\zkServer.cmd > > > >>> > > > >>> Step 4 - Start Main SOLR instance > > > >>> ========================== > > > >>> I ran the following command to start the main SOLR instance > > > >>> > > > >>> java -Djetty.port=8081 -Dhostport=8081 > > > >>> -Dbootstrap_configdir=[DATA_DIRECTORY]/solr/conf -Dnumshards=2 > > > >>> -Dzkhost=localhost:2181,localhost:2182,localhost:2183 -jar > start.jar > > > >>> > > > >>> Starts up fine. > > > >>> > > > >>> Step 5 - Start the Remaining 3 SOLR Instances > > > >>> ================================== > > > >>> I ran the following commands to start the other 3 instances from > their > > > >>> home directories: > > > >>> > > > >>> java -Djetty.port=8082 -Dhostport=8082 > > > >>> -Dzkhost=localhost:2181,localhost:2182,localhost:2183 -jar > start.jar > > > >>> > > > >>> java -Djetty.port=8083 -Dhostport=8083 > > > >>> -Dzkhost=localhost:2181,localhost:2182,localhost:2183 -jar > start.jar > > > >>> > > > >>> java -Djetty.port=8084 -Dhostport=8084 > > > >>> -Dzkhost=localhost:2181,localhost:2182,localhost:2183 -jar > start.jar > > > >>> > > > >>> All startup without issue. > > > >>> > > > >>> Step 6 - Modified solrconfig.xml to have a custom request handler > > > >>> =============================================== > > > >>> > > > >>> <requestHandler name="/update/sharepoint" startup="lazy" > > > >>> class="solr.extraction.ExtractingRequestHandler"> > > > >>> <lst name="defaults"> > > > >>> <str name="update.chain">sharepoint-pipeline</str> > > > >>> <str name="fmap.content">text</str> > > > >>> <str name="lowernames">true</str> > > > >>> <str name="uprefix">ignored</str> > > > >>> <str name="caputreAttr">true</str> > > > >>> <str name="fmap.a">links</str> > > > >>> <str name="fmap.div">ignored</str> > > > >>> </lst> > > > >>> </requestHandler> > > > >>> > > > >>> <updateRequestProcessorChain name="sharepoint-pipeline"> > > > >>> <processor > class="solr.processor.SignatureUpdateProcessorFactory"> > > > >>> <bool name="enabled">true</bool> > > > >>> <str name="signatureField">id</str> > > > >>> <bool name="owerrightDupes">true</bool> > > > >>> <str name="fields">url</str> > > > >>> <str > name="signatureClass">solr.processor.Lookup3Signature</str> > > > >>> </processor> > > > >>> <processor class="solr.LogUpdateProcessorFactory"/> > > > >>> <processor class="solr.RunUpdateProcessorFactory"/> > > > >>> </updateRequestProcessorChain> > > > >>> > > > >>> > > > >>> Hopefully this will shed some light on why my configuration is > having > > > >>> issues. > > > >>> > > > >>> Thanks for your help. > > > >>> > > > >>> Matt > > > >>> > > > >>> > > > >>> > > > >>> On Tue, Feb 28, 2012 at 8:29 PM, Mark Miller < > markrmil...@gmail.com > > > >wrote: > > > >>> > > > >>>> Hmm...this is very strange - there is nothing interesting in any > of > > > the > > > >>>> logs? > > > >>>> > > > >>>> In clusterstate.json, all of the shards have an active state? > > > >>>> > > > >>>> > > > >>>> There are quite a few of us doing exactly this setup recently, so > > > there > > > >>>> must be something we are missing here... > > > >>>> > > > >>>> Any info you can offer might help. > > > >>>> > > > >>>> - Mark > > > >>>> > > > >>>> On Feb 28, 2012, at 1:00 PM, Matthew Parker wrote: > > > >>>> > > > >>>> > Mark, > > > >>>> > > > > >>>> > I got the codebase from the 2/26/2012, and I got the same > > > inconsistent > > > >>>> > results. > > > >>>> > > > > >>>> > I have solr running on four ports 8081-8084 > > > >>>> > > > > >>>> > 8081 and 8082 are the leaders for shard 1, and shard 2, > respectively > > > >>>> > > > > >>>> > 8083 - is assigned to shard 1 > > > >>>> > 8084 - is assigned to shard 2 > > > >>>> > > > > >>>> > queries come in and sometime it seems the windows from 8081 and > 8083 > > > >>>> move > > > >>>> > responding to the query but there are no results. > > > >>>> > > > > >>>> > if the queries run on 8081/8082 or 8081/8084 then results come > back > > > >>>> ok. > > > >>>> > > > > >>>> > The query is nothing more than: q=*:* > > > >>>> > > > > >>>> > Regards, > > > >>>> > > > > >>>> > Matt > > > >>>> > > > > >>>> > > > > >>>> > On Mon, Feb 27, 2012 at 9:26 PM, Matthew Parker < > > > >>>> > mpar...@apogeeintegration.com> wrote: > > > >>>> > > > > >>>> >> I'll have to check on the commit situation. We have been > pushing > > > >>>> data from > > > >>>> >> SharePoint the last week or so. Would that somehow block the > > > >>>> documents > > > >>>> >> moving between the solr instances? > > > >>>> >> > > > >>>> >> I'll try another version tomorrow. Thanks for the suggestions. > > > >>>> >> > > > >>>> >> On Mon, Feb 27, 2012 at 5:34 PM, Mark Miller < > > > markrmil...@gmail.com > > > >>>> >wrote: > > > >>>> >> > > > >>>> >>> Hmmm...all of that looks pretty normal... > > > >>>> >>> > > > >>>> >>> Did a commit somehow fail on the other machine? When you view > the > > > >>>> stats > > > >>>> >>> for the update handler, are there a lot of pending adds for > on of > > > >>>> the > > > >>>> >>> nodes? Do the commit counts match across nodes? > > > >>>> >>> > > > >>>> >>> You can also query an individual node with distrib=false to > check > > > >>>> that. > > > >>>> >>> > > > >>>> >>> If you build is a month old, I'd honestly recommend you try > > > >>>> upgrading as > > > >>>> >>> well. > > > >>>> >>> > > > >>>> >>> - Mark > > > >>>> >>> > > > >>>> >>> On Feb 27, 2012, at 3:34 PM, Matthew Parker wrote: > > > >>>> >>> > > > >>>> >>>> Here is most of the cluster state: > > > >>>> >>>> > > > >>>> >>>> Connected to Zookeeper > > > >>>> >>>> localhost:2181, localhost: 2182, localhost:2183 > > > >>>> >>>> > > > >>>> >>>> /(v=0 children=7) "" > > > >>>> >>>> /CONFIGS(v=0, children=1) > > > >>>> >>>> /CONFIGURATION(v=0 children=25) > > > >>>> >>>> <<<<< all the configuration files, velocity info, > > > xslt, > > > >>>> etc. > > > >>>> >>>>>>>> > > > >>>> >>>> /NODE_STATES(v=0 children=4) > > > >>>> >>>> MACHINE1:8083_SOLR (v=121)"[{"shard_id":"shard1", > > > >>>> >>>> > > > >>>> > "state":"active","core":"","collection":"collection1","node_name:"..." > > > >>>> >>>> MACHINE1:8082_SOLR (v=101)"[{"shard_id":"shard2", > > > >>>> >>>> > > > >>>> > "state":"active","core":"","collection":"collection1","node_name:"..." > > > >>>> >>>> MACHINE1:8081_SOLR (v=92)"[{"shard_id":"shard1", > > > >>>> >>>> > > > >>>> > "state":"active","core":"","collection":"collection1","node_name:"..." > > > >>>> >>>> MACHINE1:8084_SOLR (v=73)"[{"shard_id":"shard2", > > > >>>> >>>> > > > >>>> > "state":"active","core":"","collection":"collection1","node_name:"..." > > > >>>> >>>> /ZOOKEEPER (v-0 children=1) > > > >>>> >>>> QUOTA(v=0) > > > >>>> >>>> > > > >>>> >>>> > > > >>>> >>> > > > >>>> > > > > /CLUSTERSTATE.JSON(V=272)"{"collection1":{"shard1":{MACHINE1:8081_solr_":{shard_id":"shard1","leader":"true","..." > > > >>>> >>>> /LIVE_NODES (v=0 children=4) > > > >>>> >>>> MACHINE1:8083_SOLR(ephemeral v=0) > > > >>>> >>>> MACHINE1:8082_SOLR(ephemeral v=0) > > > >>>> >>>> MACHINE1:8081_SOLR(ephemeral v=0) > > > >>>> >>>> MACHINE1:8084_SOLR(ephemeral v=0) > > > >>>> >>>> /COLLECTIONS (v=1 children=1) > > > >>>> >>>> COLLECTION1(v=0 > children=2)"{"configName":"configuration1"}" > > > >>>> >>>> LEADER_ELECT(v=0 children=2) > > > >>>> >>>> SHARD1(V=0 children=1) > > > >>>> >>>> ELECTION(v=0 children=2) > > > >>>> >>>> > > > >>>> >>>> 87186203314552835-MACHINE1:8081_SOLR_-N_0000000096(ephemeral > v=0) > > > >>>> >>>> > > > >>>> >>>> 87186203314552836-MACHINE1:8083_SOLR_-N_0000000084(ephemeral > v=0) > > > >>>> >>>> SHARD2(v=0 children=1) > > > >>>> >>>> ELECTION(v=0 children=2) > > > >>>> >>>> > > > >>>> >>>> 231301391392833539-MACHINE1:8084_SOLR_-N_0000000085(ephemeral > > > v=0) > > > >>>> >>>> > > > >>>> >>>> 159243797356740611-MACHINE1:8082_SOLR_-N_0000000084(ephemeral > > > v=0) > > > >>>> >>>> LEADERS (v=0 children=2) > > > >>>> >>>> SHARD1 (ephemeral > > > >>>> >>>> v=0)"{"core":"","node_name":"MACHINE1:8081_solr","base_url":" > > > >>>> >>>> http://MACHINE1:8081/solr"}" > > > >>>> >>>> SHARD2 (ephemeral > > > >>>> >>>> v=0)"{"core":"","node_name":"MACHINE1:8082_solr","base_url":" > > > >>>> >>>> http://MACHINE1:8082/solr"}" > > > >>>> >>>> /OVERSEER_ELECT (v=0 children=2) > > > >>>> >>>> ELECTION (v=0 children=4) > > > >>>> >>>> > > > >>>> 231301391392833539-MACHINE1:8084_SOLR_-N_0000000251(ephemeral > > > >>>> >>> v=0) > > > >>>> >>>> > > > 87186203314552835-MACHINE1:8081_SOLR_-N_0000000248(ephemeral > > > >>>> >>> v=0) > > > >>>> >>>> > > > >>>> 159243797356740611-MACHINE1:8082_SOLR_-N_0000000250(ephemeral > > > >>>> >>> v=0) > > > >>>> >>>> > > > 87186203314552836-MACHINE1:8083_SOLR_-N_0000000249(ephemeral > > > >>>> >>> v=0) > > > >>>> >>>> LEADER (emphemeral > > > >>>> >>>> > v=0)"{"id":"87186203314552835-MACHINE1:8081_solr-n_000000248"}" > > > >>>> >>>> > > > >>>> >>>> > > > >>>> >>>> > > > >>>> >>>> On Mon, Feb 27, 2012 at 2:47 PM, Mark Miller < > > > >>>> markrmil...@gmail.com> > > > >>>> >>> wrote: > > > >>>> >>>> > > > >>>> >>>>> > > > >>>> >>>>> On Feb 27, 2012, at 2:22 PM, Matthew Parker wrote: > > > >>>> >>>>> > > > >>>> >>>>>> Thanks for your reply Mark. > > > >>>> >>>>>> > > > >>>> >>>>>> I believe the build was towards the begining of the month. > The > > > >>>> >>>>>> solr.spec.version is 4.0.0.2012.01.10.38.09 > > > >>>> >>>>>> > > > >>>> >>>>>> I cannot access the clusterstate.json contents. I clicked > on > > > it a > > > >>>> >>> couple > > > >>>> >>>>> of > > > >>>> >>>>>> times, but nothing happens. Is that stored on disk > somewhere? > > > >>>> >>>>> > > > >>>> >>>>> Are you using the new admin UI? That has recently been > updated > > > to > > > >>>> work > > > >>>> >>>>> better with cloud - it had some troubles not too long ago. > If > > > you > > > >>>> are, > > > >>>> >>> you > > > >>>> >>>>> should trying using the old admin UI's zookeeper page - that > > > >>>> should > > > >>>> >>> show > > > >>>> >>>>> the cluster state. > > > >>>> >>>>> > > > >>>> >>>>> That being said, there has been a lot of bug fixes over the > past > > > >>>> month > > > >>>> >>> - > > > >>>> >>>>> so you may just want to update to a recent version. > > > >>>> >>>>> > > > >>>> >>>>>> > > > >>>> >>>>>> I configured a custom request handler to calculate an > unique > > > >>>> document > > > >>>> >>> id > > > >>>> >>>>>> based on the file's url. > > > >>>> >>>>>> > > > >>>> >>>>>> On Mon, Feb 27, 2012 at 1:13 PM, Mark Miller < > > > >>>> markrmil...@gmail.com> > > > >>>> >>>>> wrote: > > > >>>> >>>>>> > > > >>>> >>>>>>> Hey Matt - is your build recent? > > > >>>> >>>>>>> > > > >>>> >>>>>>> Can you visit the cloud/zookeeper page in the admin and > send > > > the > > > >>>> >>>>> contents > > > >>>> >>>>>>> of the clusterstate.json node? > > > >>>> >>>>>>> > > > >>>> >>>>>>> Are you using a custom index chain or anything out of the > > > >>>> ordinary? > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> - Mark > > > >>>> >>>>>>> > > > >>>> >>>>>>> On Feb 27, 2012, at 12:26 PM, Matthew Parker wrote: > > > >>>> >>>>>>> > > > >>>> >>>>>>>> TWIMC: > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> Environment > > > >>>> >>>>>>>> ========= > > > >>>> >>>>>>>> Apache SOLR rev-1236154 > > > >>>> >>>>>>>> Apache Zookeeper 3.3.4 > > > >>>> >>>>>>>> Windows 7 > > > >>>> >>>>>>>> JDK 1.6.0_23.b05 > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> I have built a SOLR Cloud instance with 4 nodes using the > > > >>>> embeded > > > >>>> >>> Jetty > > > >>>> >>>>>>>> servers. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> I created a 3 node zookeeper ensemble to manage the solr > > > >>>> >>> configuration > > > >>>> >>>>>>> data. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> All the instances run on one server so I've had to move > ports > > > >>>> around > > > >>>> >>>>> for > > > >>>> >>>>>>>> the various applications. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> I start the 3 zookeeper nodes. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> I started the first instance of solr cloud with the > parameter > > > >>>> to > > > >>>> >>> have > > > >>>> >>>>> two > > > >>>> >>>>>>>> shards. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> The start the remaining 3 solr nodes. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> The system comes up fine. No errors thrown. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> I can view the solr cloud console and I can see the SOLR > > > >>>> >>> configuration > > > >>>> >>>>>>>> files managed by ZooKeeper. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> I published data into the SOLR Cloud instances from > > > SharePoint > > > >>>> using > > > >>>> >>>>>>> Apache > > > >>>> >>>>>>>> Manifold 0.4-incubating. Manifold is setup to publish the > > > data > > > >>>> into > > > >>>> >>>>>>>> collection1, which is the only collection defined in the > > > >>>> cluster. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> When I query the data from collection1 as per the solr > wiki, > > > >>>> the > > > >>>> >>>>> results > > > >>>> >>>>>>>> are inconsistent. Sometimes all the results are there, > other > > > >>>> times > > > >>>> >>>>>>> nothing > > > >>>> >>>>>>>> comes back at all. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> It seems to be having an issue auto replicating the data > > > >>>> across the > > > >>>> >>>>>>> cloud. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> Is there some specific setting I might have missed? Based > > > upon > > > >>>> what > > > >>>> >>> I > > > >>>> >>>>>>> read, > > > >>>> >>>>>>>> I thought that SOLR cloud would take care of > distributing and > > > >>>> >>>>> replicating > > > >>>> >>>>>>>> the data automatically. Do you have to tell it what > shard to > > > >>>> publish > > > >>>> >>>>> the > > > >>>> >>>>>>>> data into as well? > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> Any help would be appreciated. > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> Thanks, > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> Matt > > > >>>> >>>>>>>> > > > >>>> >>>>>>>> ------------------------------ > > > >>>> >>>>>>>> This e-mail and any files transmitted with it may be > > > >>>> proprietary. > > > >>>> >>>>>>> Please note that any views or opinions presented in this > > > e-mail > > > >>>> are > > > >>>> >>>>> solely > > > >>>> >>>>>>> those of the author and do not necessarily represent > those of > > > >>>> Apogee > > > >>>> >>>>>>> Integration. > > > >>>> >>>>>>> > > > >>>> >>>>>>> - Mark Miller > > > >>>> >>>>>>> lucidimagination.com > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>>> > > > >>>> >>>>>> > > > >>>> >>>>>> Matt > > > >>>> >>>>>> > > > >>>> >>>>>> ------------------------------ > > > >>>> >>>>>> This e-mail and any files transmitted with it may be > > > proprietary. > > > >>>> >>>>> Please note that any views or opinions presented in this > e-mail > > > >>>> are > > > >>>> >>> solely > > > >>>> >>>>> those of the author and do not necessarily represent those > of > > > >>>> Apogee > > > >>>> >>>>> Integration. > > > >>>> >>>>> > > > >>>> >>>>> - Mark Miller > > > >>>> >>>>> lucidimagination.com > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>>> > > > >>>> >>>> > > > >>>> >>>> ------------------------------ > > > >>>> >>>> This e-mail and any files transmitted with it may be > proprietary. > > > >>>> >>> Please note that any views or opinions presented in this > e-mail > > > are > > > >>>> solely > > > >>>> >>> those of the author and do not necessarily represent those of > > > Apogee > > > >>>> >>> Integration. > > > >>>> >>> > > > >>>> >>> - Mark Miller > > > >>>> >>> lucidimagination.com > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >>> > > > >>>> >> > > > >>>> > > > > >>>> > ------------------------------ > > > >>>> > This e-mail and any files transmitted with it may be > proprietary. > > > >>>> Please note that any views or opinions presented in this e-mail > are > > > solely > > > >>>> those of the author and do not necessarily represent those of > Apogee > > > >>>> Integration. > > > >>>> > > > >>>> - Mark Miller > > > >>>> lucidimagination.com > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>>> > > > >>> > > > >> > > > > > > > > > > ------------------------------ > > > This e-mail and any files transmitted with it may be proprietary. > Please > > > note that any views or opinions presented in this e-mail are solely > those > > > of the author and do not necessarily represent those of Apogee > Integration. > > > > > > > > > > > -- > > - Mark > > > > http://www.lucidimagination.com > > > > ------------------------------ > > This e-mail and any files transmitted with it may be proprietary. > Please note that any views or opinions presented in this e-mail are solely > those of the author and do not necessarily represent those of Apogee > Integration. > > > > > > - Mark Miller > lucidimagination.com > > > > > > > > > > > > ------------------------------ This e-mail and any files transmitted with it may be proprietary. Please note that any views or opinions presented in this e-mail are solely those of the author and do not necessarily represent those of Apogee Integration.