Re: "Classic" 4.2 master-slave replication not completing

2013-06-27 Thread Neal Ensor
Okay, I have done this (updated to 4.3.1 across master and four slaves; one
of these is my own PC for experiments, it is not being accessed by clients).

Just had a minor replication this morning, and all three slaves are "stuck"
again.  Replication supposedly started at 8:40, ended 30 seconds later or
so (on my local PC, set up identically to the other three slaves).  The
three slaves will NOT complete the roll-over to the new index.  All three
index folders have a write.lock and latest files are dated 8:40am (now it
is 8:54am, with no further activity in the index folders).  There exists an
"index.2013062708461" (or some variation thereof) in all three slaves'
data folder.

The seemingly-relevant thread dump of a "snappuller" thread on each of
these slaves:

   - sun.misc.Unsafe.park(Native Method)
   - java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
   -
   
java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:811)
   -
   
java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:969)
   -
   
java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1281)
   - java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:218)
   - java.util.concurrent.FutureTask.get(FutureTask.java:83)
   -
   
org.apache.solr.handler.SnapPuller.openNewWriterAndSearcher(SnapPuller.java:631)
   -
   org.apache.solr.handler.SnapPuller.fetchLatestIndex(SnapPuller.java:446)
   -
   
org.apache.solr.handler.ReplicationHandler.doFetch(ReplicationHandler.java:317)
   - org.apache.solr.handler.SnapPuller$1.run(SnapPuller.java:223)
   -
   java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
   -
   java.util.concurrent.FutureTask$Sync.innerRunAndReset(FutureTask.java:317)
   - java.util.concurrent.FutureTask.runAndReset(FutureTask.java:150)
   -
   
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$101(ScheduledThreadPoolExecutor.java:98)
   -
   
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.runPeriodic(ScheduledThreadPoolExecutor.java:180)
   -
   
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:204)
   -
   
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
   -
   
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
   - java.lang.Thread.run(Thread.java:662)


Here they sit.  My local PC "slave" replicated very quickly, switched over
to the new generation (206) immediately.  I am not sure why the three
slaves are dragging on this.  If there's any configuration elements or
other details you need, please let me know.  I can manually "kick" them by
reloading the core from the admin pages, but obviously I would like this to
be a hands-off process.  Any help is greatly appreciated; this has been
bugging me for some time now.



On Mon, Jun 24, 2013 at 9:34 AM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:

> A bunch of replication related issues were fixed in 4.2.1 so you're
> better off upgrading to 4.2.1 or later (4.3.1 is the latest release).
>
> On Mon, Jun 24, 2013 at 6:55 PM, Neal Ensor  wrote:
> > As a bit of background, we run a setup (coming from 3.6.1 to 4.2
> relatively
> > recently) with a single master receiving updates with three slaves
> pulling
> > changes in.  Our index is around 5 million documents, around 26GB in size
> > total.
> >
> > The situation I'm seeing is this:  occasionally we update the master, and
> > replication begins on the three slaves, seems to proceed normally until
> it
> > hits the end.  At that point, it "sticks"; there's no messages going on
> in
> > the logs, nothing on the admin page seems to be happening.  I sit there
> for
> > sometimes upwards of 30 minutes, seeing no further activity in the index
> > folder(s).   After a while, I go to the core admin page and manually
> reload
> > the core, which "catches it up".  It seems like the index readers /
> writers
> > are not releasing the index otherwise?  The configuration is set to
> reopen;
> > very occasionally this situation actually fixes itself after a longish
> > period of time, but it seems very annoying.
> >
> > I had at first suspected this to be due to our underlying shared (SAN)
> > storage, so we installed SSDs in all three slave machines, and moved the
> > entire indexes to those.  It did not seem to affect this issue at all
> > (additionally, I didn't really see the expected performance boost, but
> > that's a separate issue entirely).
> >
> > Any ideas?  Any configuration details I might share/reconfigure?  Any
> > suggestions are appreciated. I could also upgrade to the later 4.3+
> > versions, if that might help.
> >
> > Thanks!
> >
> > Neal Ensor
> > nen...@gmail.com
>
>
>
> --
> Regards,
> Shalin Shekhar Mangar.
>


Re: "Classic" 4.2 master-slave replication not completing

2013-07-01 Thread Neal Ensor
is it conceivable that there's too much traffic, causing Solr to stall
re-opening the searcher (thus releasing to the new index)?  I'm grasping at
straws, and this is beginning to bug me a lot.  The traffic logs wouldn't
seem to support this (apart from periodic health-check pings, the load is
distributed fairly evenly across 3 slaves by a load-balancer tool).  After
35+ minutes this morning, none of the three successfully "unstuck", and had
to be manually core-reloaded.

Is there perhaps a configuration element I'm overlooking that might make
solr a bit less "friendly" about it, and just dump the searchers/reopen
when replication completes?

As a side note, I'm getting really frustrated with trying to get log4j
logging on 4.3.1 set up; my tomcat container persists in complaining that
it cannot find log4j.properties, when I've put it in the WEB-INF/classes of
the war file, have SLF4j libraries AND log4j at the shared container "lib"
level, and log4j.debug turned on.  I can't find any excuses why it cannot
seem to locate the configuration.

Any suggestions or pointers would be greatly appreciated.  Thanks!


On Thu, Jun 27, 2013 at 10:35 AM, Mark Miller  wrote:

> Odd - looks like it's stuck waiting to be notified that a new searcher is
> ready.
>
> - Mark
>
> On Jun 27, 2013, at 8:58 AM, Neal Ensor  wrote:
>
> > Okay, I have done this (updated to 4.3.1 across master and four slaves;
> one
> > of these is my own PC for experiments, it is not being accessed by
> clients).
> >
> > Just had a minor replication this morning, and all three slaves are
> "stuck"
> > again.  Replication supposedly started at 8:40, ended 30 seconds later or
> > so (on my local PC, set up identically to the other three slaves).  The
> > three slaves will NOT complete the roll-over to the new index.  All three
> > index folders have a write.lock and latest files are dated 8:40am (now it
> > is 8:54am, with no further activity in the index folders).  There exists
> an
> > "index.2013062708461" (or some variation thereof) in all three
> slaves'
> > data folder.
> >
> > The seemingly-relevant thread dump of a "snappuller" thread on each of
> > these slaves:
> >
> >   - sun.misc.Unsafe.park(Native Method)
> >   - java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
> >   -
> >
> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:811)
> >   -
> >
> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:969)
> >   -
> >
> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1281)
> >   - java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:218)
> >   - java.util.concurrent.FutureTask.get(FutureTask.java:83)
> >   -
> >
> org.apache.solr.handler.SnapPuller.openNewWriterAndSearcher(SnapPuller.java:631)
> >   -
> >
> org.apache.solr.handler.SnapPuller.fetchLatestIndex(SnapPuller.java:446)
> >   -
> >
> org.apache.solr.handler.ReplicationHandler.doFetch(ReplicationHandler.java:317)
> >   - org.apache.solr.handler.SnapPuller$1.run(SnapPuller.java:223)
> >   -
> >   java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
> >   -
> >
> java.util.concurrent.FutureTask$Sync.innerRunAndReset(FutureTask.java:317)
> >   - java.util.concurrent.FutureTask.runAndReset(FutureTask.java:150)
> >   -
> >
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$101(ScheduledThreadPoolExecutor.java:98)
> >   -
> >
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.runPeriodic(ScheduledThreadPoolExecutor.java:180)
> >   -
> >
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:204)
> >   -
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
> >   -
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
> >   - java.lang.Thread.run(Thread.java:662)
> >
> >
> > Here they sit.  My local PC "slave" replicated very quickly, switched
> over
> > to the new generation (206) immediately.  I am not sure why the three
> > slaves are dragging on this.  If there's any configuration elements or
> > other details you need, please let me know.  I can manually "kick" them
> by
> > reloading the core from the admin pages, but obviously I would like this
> to
> &g

Re: "Classic" 4.2 master-slave replication not completing

2013-07-03 Thread Neal Ensor
Posted the solr config up as http://apaste.info/4eKC (hope that works).
 Note that this is largely a hold-over from upgrades of previous solr
versions, there may be lots of cruft left over.  If it's advisable to do
so, I would certainly be open to starting from scratch with a 4.3+ example
configuration.

I will take up the logging portions you listed there, to see where I might
have missed a library or configuration point, thanks.


On Mon, Jul 1, 2013 at 4:04 PM, Shawn Heisey  wrote:

> On 7/1/2013 1:07 PM, Neal Ensor wrote:
>
>> is it conceivable that there's too much traffic, causing Solr to stall
>> re-opening the searcher (thus releasing to the new index)?  I'm grasping
>> at
>> straws, and this is beginning to bug me a lot.  The traffic logs wouldn't
>> seem to support this (apart from periodic health-check pings, the load is
>> distributed fairly evenly across 3 slaves by a load-balancer tool).  After
>> 35+ minutes this morning, none of the three successfully "unstuck", and
>> had
>> to be manually core-reloaded.
>>
>> Is there perhaps a configuration element I'm overlooking that might make
>> solr a bit less "friendly" about it, and just dump the searchers/reopen
>> when replication completes?
>>
>
> Can you share your solrconfig.xml file, someplace like http://apaste.info?
>  Please be sure to choose the correct file type ... on that website it is
> (X)HTML for an XML file.
>
>
>  As a side note, I'm getting really frustrated with trying to get log4j
>> logging on 4.3.1 set up; my tomcat container persists in complaining that
>> it cannot find log4j.properties, when I've put it in the WEB-INF/classes
>> of
>> the war file, have SLF4j libraries AND log4j at the shared container "lib"
>> level, and log4j.debug turned on.  I can't find any excuses why it cannot
>> seem to locate the configuration.
>>
>
> The wiki is still down for maintenance, so below is a relevant section of
> the SolrLogging wiki page extracted from Google Cache.  When it comes back
> up, you can find it at this URL:
>
> http://wiki.apache.org/solr/**SolrLogging#Switching_from_**
> Log4J_back_to_JUL_.28java.**util.logging.29<http://wiki.apache.org/solr/SolrLogging#Switching_from_Log4J_back_to_JUL_.28java.util.logging.29>
>
> =
> The example logging setup takes over the configuration of Solr logging,
> which prevents the container from controlling where logs go. Users of
> containers other than the included Jetty (Tomcat in particular) may be
> accustomed to doing the logging configuration in the container. If you want
> to switch back to java.util.logging so this is once again possible, here's
> what to do. These steps apply to the example/lib/ext directory in the Solr
> example, or to your container's lib directory as mentioned in the previous
> section. These steps also assume that the slf4j version is 1.6.6, which
> comes with Solr4.3. Newer versions may use a different slf4j version. As of
> May 2013, you can use a newer SLF4J version with no trouble, but be aware
> that all slf4j components in your classpath must be the same version.
>
> Download slf4j version 1.6.6 (the version used in Solr4.3.x).
> http://www.slf4j.org/dist/**slf4j-1.6.6.zip<http://www.slf4j.org/dist/slf4j-1.6.6.zip>
> Unpack the slf4j archive.
> Delete these JARs from your lib folder: slf4j-log4j12-1.6.6.jar,
> jul-to-slf4j-1.6.6.jar, log4j-1.2.16.jar
> Add these JARs to your lib folder (from slf4j zip):
> slf4j-jdk14-1.6.6.jar, log4j-over-slf4j-1.6.6.jar
> Use your old logging.properties
> =
>
> Thanks,
> Shawn
>
>


Re: "Classic" 4.2 master-slave replication not completing

2013-07-03 Thread Neal Ensor
I'll give that a shot, thanks!


On Wed, Jul 3, 2013 at 12:28 PM, Shawn Heisey  wrote:

> On 7/3/2013 9:29 AM, Neal Ensor wrote:
>
>> Posted the solr config up as http://apaste.info/4eKC (hope that works).
>>   Note that this is largely a hold-over from upgrades of previous solr
>> versions, there may be lots of cruft left over.  If it's advisable to do
>> so, I would certainly be open to starting from scratch with a 4.3+ example
>> configuration.
>>
>> I will take up the logging portions you listed there, to see where I might
>> have missed a library or configuration point, thanks.
>>
>
> You have a LOT of auto warming in your cache configuration.  Unless your
> filter queries are really super simple, a more realistic autowarmCount for
> the filterCache is low double digits or even single digit.  If they are
> super simple, a number like 64 might be OK.  For the queryResultCache, I
> don't think I would go above a few hundred at the absolute most.  You
> probably don't need overall cache sizes that large, but that shouldn't
> cause any problems, unless it is pushing the limits of your java heap.
>
> The high numbers that you have mean that every commit with openSearcher
> set to true (the default) will result in Solr running 4608 queries against
> the new index before it can go online.  If both master and slave servers
> have similar config, it will happen on both.  As you might imagine, this
> many queries can take a very long time, which can lead to a whole cascade
> of other problems, and I wouldn't be surprised to learn that replication is
> affected.
>
> Try reducing the autowarmCount values drastically and see what happens.
>
> Thanks,
> Shawn
>
>


Question about filter query: "half" of my index is slower than the other?

2013-08-09 Thread Neal Ensor
So, I have an oddball question I have been battling with in the last day or
two.

I have an 8 million document solr index, roughly divided down the middle by
an identifying "product" value, one of two distinct values.  The documents
in both "sides" are very similar, with stored text fields, etc.  I have two
nearly identical request handlers, one for each "side".

When I perform very similar queries on either "side" for random phrases,
requesting 500 rows with highlighting on titles and summaries, I get very
different results.  One "side" consistently returns results in around 1-2
seconds, whereas the other one consistently returns in 6-10 seconds.  I
don't see any reason why it's worse; each run of queries is deliberately
randomized to avoid caches getting in the way.  Each test query returns the
full first 500 in most cases.

My filter query cache configuration looks like:



(desperately trying to increase it, hoping this would help).  The other
caches are quite small; the use cases the customer is dealing with don't
involve much in the way of paging, just returning a large initial set with
highlighting in the shortest time.

I'm trying to optimize this down so the disparity between the two "halves"
is not so dramatic.  Is there any optimizations or things I should be
looking for to tune?  Is it just the "way it is"?  I've tried to argue to
decrease the return set size, turn off highlighting, etc., but these seem
to be out of the question.  I would at least like some concrete reason why
one filter query would be so relatively out of whack than the other, given
the document ranges are very nearly half (3.8 million vs. 4.0 million in
the slower side).

Any pointers or suggestions would be appreciated.  Thanks in advance.

Neal Ensor
nen...@gmail.com


Re: Question about filter query: "half" of my index is slower than the other?

2013-08-09 Thread Neal Ensor
It seems (from observation only) that most of the documents on both sides
of this equation have the same "weights".  I don't see any wide swaths of
unpopulated fields on the "good" side.  Just wondering if there's some
caching involved that I'm missing here, or something I can balance out
better...


On Fri, Aug 9, 2013 at 11:39 AM, Raymond Wiker  wrote:

> On Aug 9, 2013, at 17:36 , Neal Ensor  wrote:
> > So, I have an oddball question I have been battling with in the last day
> or
> > two.
> >
> > I have an 8 million document solr index, roughly divided down the middle
> by
> > an identifying "product" value, one of two distinct values.  The
> documents
> > in both "sides" are very similar, with stored text fields, etc.  I have
> two
> > nearly identical request handlers, one for each "side".
> >
> > When I perform very similar queries on either "side" for random phrases,
> > requesting 500 rows with highlighting on titles and summaries, I get very
> > different results.  One "side" consistently returns results in around 1-2
> > seconds, whereas the other one consistently returns in 6-10 seconds.  I
> > don't see any reason why it's worse; each run of queries is deliberately
> > randomized to avoid caches getting in the way.  Each test query returns
> the
> > full first 500 in most cases.
> >
> > My filter query cache configuration looks like:
> >
> >  > size="75"
> > initialSize="1"
> > autowarmCount="0"/>
> >
> > (desperately trying to increase it, hoping this would help).  The other
> > caches are quite small; the use cases the customer is dealing with don't
> > involve much in the way of paging, just returning a large initial set
> with
> > highlighting in the shortest time.
> >
> > I'm trying to optimize this down so the disparity between the two
> "halves"
> > is not so dramatic.  Is there any optimizations or things I should be
> > looking for to tune?  Is it just the "way it is"?  I've tried to argue to
> > decrease the return set size, turn off highlighting, etc., but these seem
> > to be out of the question.  I would at least like some concrete reason
> why
> > one filter query would be so relatively out of whack than the other,
> given
> > the document ranges are very nearly half (3.8 million vs. 4.0 million in
> > the slower side).
> >
> > Any pointers or suggestions would be appreciated.  Thanks in advance.
> >
> > Neal Ensor
> > nen...@gmail.com
>
> Does one side have mcuh more data in one of the fields that is being
> returned?


SOLR 4.2, slaves replicating reporting higher version number than master

2013-09-16 Thread Neal Ensor
Having a strange intermittent issue with my 1 master, 3 slave solr 4.2
setup.  On occasion, after indexing the master and replicating across the
three slaves, each slave will start reporting they are one generation ahead
(525 vs. 524 on the master) and thus out of sync.  Replication runs appear
to do nothing, and it seems to not really be affecting performance, it's
just tickling my admin nerves.

Any suggestions of what to look at?  Just upgrade solr perhaps?  4.2 might
be getting rather old...


Re: SOLR 4.2, slaves replicating reporting higher version number than master

2013-09-16 Thread Neal Ensor
Looks much like what I'm encountering.  Guessing that will go away once I
update solr, just wanted to make sure it wasn't a real bug.  Entirely
possible we are getting some "empty commits" given the nature of the index
maintenance.  Thanks for the pointer!


On Mon, Sep 16, 2013 at 2:00 PM, Chris Hostetter
wrote:

>
> Sounds like perhaps you are getting confused by this...
>
> https://issues.apache.org/jira/browse/SOLR-4661
>
> ...if that is the situation then it's not a bug you need to worry about,
> just a confusion in how the ReplicaitonHandler reports it's stats -- the
> newer UI makes it more clear what numbers you are looking at.
>
> If that doesn't looke like the problem you are seeing, then more detail on
> how to reproduce what you are seeing would be helpful (replicaiton
> configs, logs from amster & slave, etc...)
>
>
> -Hoss
>


"Classic" 4.2 master-slave replication not completing

2013-06-24 Thread Neal Ensor
As a bit of background, we run a setup (coming from 3.6.1 to 4.2 relatively
recently) with a single master receiving updates with three slaves pulling
changes in.  Our index is around 5 million documents, around 26GB in size
total.

The situation I'm seeing is this:  occasionally we update the master, and
replication begins on the three slaves, seems to proceed normally until it
hits the end.  At that point, it "sticks"; there's no messages going on in
the logs, nothing on the admin page seems to be happening.  I sit there for
sometimes upwards of 30 minutes, seeing no further activity in the index
folder(s).   After a while, I go to the core admin page and manually reload
the core, which "catches it up".  It seems like the index readers / writers
are not releasing the index otherwise?  The configuration is set to reopen;
very occasionally this situation actually fixes itself after a longish
period of time, but it seems very annoying.

I had at first suspected this to be due to our underlying shared (SAN)
storage, so we installed SSDs in all three slave machines, and moved the
entire indexes to those.  It did not seem to affect this issue at all
(additionally, I didn't really see the expected performance boost, but
that's a separate issue entirely).

Any ideas?  Any configuration details I might share/reconfigure?  Any
suggestions are appreciated. I could also upgrade to the later 4.3+
versions, if that might help.

Thanks!

Neal Ensor
nen...@gmail.com


SOLR 4.1 and Tomcat 6 class loader issue (older lucene implementation in common lib)

2013-02-26 Thread Neal Ensor
So, I have had great success as a SOLR user so far, love its ease of
set up.  However, I have run into this problem (and this may be more
of a Tomcat error; if so, please point me in that direction).

My tomcat application containers have an older lucene library in the
common library "lib" folder (2.9) to support some deployed legacy
searching applications.  The SOLR I'm trying to install in this
container has in its .war file WEB-INF/lib all the 4.1 lucene
libraries, which if I read the class loader documentation correctly,
should override the shared library in tomcat.  In fact, when I deploy
it, this is the case:  the dashboard shows using 4.1.0 lucene
implementation.  However, if I stop and start the application while
the container is running, it CHANGES to 2.9!

This mysterious switcheroo caused my slaves to fail after replicating
mysteriously over the weekend, as they had been bounced (and thus 2.9)
and had received a 4.1 lucene index from the master, and promptly
failed hard.  Watching one interactively showed it replicating most of
an index, then getting an "invalid master" error about 75% in, then
going down.

Any clues as to why this might be?  Should Tomcat be ignoring
delegation here, or is it in error?  My solution so far has been to
move my solr installations off onto "pristine" containers without the
lucene library anywhere except the .war file itself.  This presents
slight problems in our server farm arrangements, and I'd like to not
have to "special case" it if possible.  Any pointers would be great
(even if its to point me at Tomcat's mailing list. :) )

Thanks!

Neal Ensor
nen...@gmail.com


SOLR 4.2 SNAPSHOT "There exists no core with name x"

2013-03-01 Thread Neal Ensor
Just pulled down the latest SVN and built it in place to try and
upgrade from 4.1 (having replication issues).  However, bringing it up
with my current core (copied) presents me with the admin page, but an
empty drop-down to select specific cores.  If I go to "Core Admin" I
can see the core, I can issue queries against it, but I cannot go to
its statistics, interactive query page, etc. at all.  Attempting to
access the "old" core admin stuff via direct URL results in the pop-up
stating "There exists no core with name myname", even though I can
clearly see it there.

I even created a brand new core via the core admin interface, and it
cannot be seen either.  Restarted the container, examined logs, etc.
to no avail.  Am I missing something obvious here?  Any serious
deviations between 4.1 and 4.2 (trunk) configuration I need to update?
 Using tomcat 6.0.36 as a container, freshly-built
solr-4.2-SNAPSHOT.war hot off the compiler.

Neal Ensor
nen...@gmail.com


Re: SOLR 4.2 SNAPSHOT "There exists no core with name x"

2013-03-01 Thread Neal Ensor
Using Chrome (latest) on Mac OSX 10.8.2.  In 4.1.0, I accessed cores
via something like:

http://machine:port/solr/#/corename

and got to the Ping, Query, Schema, etc.  I attempted a similar URL
with my local installation and got the error I mentioned.  I have one
core locally named "master", and made a new one named "testcore" via
the Cores Admin.  Neither could be accessed directly (but could be
searched via the select query URL) yet came up on the Cores Admin page
fine.  The drop-down under the main menu remains empty.

Now that you mentioned browser versions, I JUST tried it using Safari,
and it actually came up; there might be some java script thing that's
giving Chrome indigestion perhaps.  Looking at the Javascript Console
in Chrome I don't see any errors.  The apparently relevant portion of
the source on that page looks like

 

  
  

  

empty.  If I can help with any other information, please let me know.
Again, it appears to work on Safari fine hitting the same container,
so must be something Chrome-specific (perhaps something I have
disabled?)


On Fri, Mar 1, 2013 at 1:25 PM, Stefan Matheis  wrote:
> Hey Neal
>
> We changed the Navigation after 4.1 was released 
> (https://issues.apache.org/jira/browse/SOLR-4284) But you're the first one 
> reporting problems with this change (if it's related).
>
> 1) What Browser on which OS are you using?
> 2) And what is/are the Name of your core(s)?
> 3) When you're talking about the "old" core admin, what is the url you're 
> trying to access?
>
> Perhaps you can as well provide Screenshots of the Core-Admin and the 
> Drop-Down (left side in the Navigation) while it's expanded?
>
> Stefan
>
>
> On Friday, March 1, 2013 at 6:57 PM, Neal Ensor wrote:
>
>> Just pulled down the latest SVN and built it in place to try and
>> upgrade from 4.1 (having replication issues). However, bringing it up
>> with my current core (copied) presents me with the admin page, but an
>> empty drop-down to select specific cores. If I go to "Core Admin" I
>> can see the core, I can issue queries against it, but I cannot go to
>> its statistics, interactive query page, etc. at all. Attempting to
>> access the "old" core admin stuff via direct URL results in the pop-up
>> stating "There exists no core with name myname", even though I can
>> clearly see it there.
>>
>> I even created a brand new core via the core admin interface, and it
>> cannot be seen either. Restarted the container, examined logs, etc.
>> to no avail. Am I missing something obvious here? Any serious
>> deviations between 4.1 and 4.2 (trunk) configuration I need to update?
>> Using tomcat 6.0.36 as a container, freshly-built
>> solr-4.2-SNAPSHOT.war hot off the compiler.
>>
>> Neal Ensor
>> nen...@gmail.com (mailto:nen...@gmail.com)
>
>
>


Re: SOLR 4.2 SNAPSHOT "There exists no core with name x"

2013-03-04 Thread Neal Ensor
Actually, just updated Chrome this morning, and it all appears to
work.  Flushed cache as well, so could be part of that.  All's well
that ends well I suppose.

neal

On Mon, Mar 4, 2013 at 4:44 AM, Jens Grivolla  wrote:
> On 03/01/2013 07:46 PM, Neal Ensor wrote:
>>
>> Again, it appears to work on Safari fine hitting the same container,
>> so must be something Chrome-specific (perhaps something I have
>> disabled?)
>
>
> This sounds like it might just be a browser cache issue (if you used Chrome
> to access the same URL previously with the old Solr version installed). It
> might just not be refreshing everything.
>
> Jens
>


Re: SOLR 4.2 SNAPSHOT "There exists no core with name x"

2013-03-04 Thread Neal Ensor
Confirmed it was indeed caching, as I just updated my live master from
4.1 to the 4.2 snap shot and got the empty drop-down; cleared cache
and reloaded, and working.

On Mon, Mar 4, 2013 at 10:03 AM, Stefan Matheis
 wrote:
> Thanks Jens! Didn't think about caching .. :/
>
> Perhaps we should change the requests in favor of 
> https://issues.apache.org/jira/browse/SOLR-4311 to avoid any caching at the 
> UI? Results maybe in a few more (real) requests but i guess that would be 
> okay?
>
> Stefan
>
>
> On Monday, March 4, 2013 at 2:21 PM, Neal Ensor wrote:
>
>> Actually, just updated Chrome this morning, and it all appears to
>> work. Flushed cache as well, so could be part of that. All's well
>> that ends well I suppose.
>>
>> neal
>>
>> On Mon, Mar 4, 2013 at 4:44 AM, Jens Grivolla > (mailto:j+...@grivolla.net)> wrote:
>> > On 03/01/2013 07:46 PM, Neal Ensor wrote:
>> > >
>> > > Again, it appears to work on Safari fine hitting the same container,
>> > > so must be something Chrome-specific (perhaps something I have
>> > > disabled?)
>> >
>> >
>> >
>> >
>> > This sounds like it might just be a browser cache issue (if you used Chrome
>> > to access the same URL previously with the old Solr version installed). It
>> > might just not be refreshing everything.
>> >
>> > Jens
>
>


SOLR 4.2 (snapshot, March 1, 2013): replication not "finishing"?

2013-03-07 Thread Neal Ensor
I'm getting intermittent issues with replication in my current
arrangement:  one master, 3 slaves; all the same SOLR version/war file
deployment.

I update the master, which kicks off replication across the other
three; however, they never seem to "finish".  In the data/ folders I
get an empty index.timestamp folder, the admin page for replication
shows it "stuck" pulling a file (no progress shown, just constant
refresh).  The index never changes over (always claiming to be
out-of-date).  Abort messages are ignored, both from the admin console
and through a curl "abortfetch" request to the slaves.  The searcher
is still responsive, I can query, but the master's changes are of
course not there.

If I kill my container (tomcat 6) and start it back up, magically the
replication has "finished" and the slave is up-to-date.  This sort of
leads me to believe something isn't finalizing the change over and
opening a new searcher on the index (looks like the "main" index is
actually being updated, but a close/open, or reopen, is not
happening?)

The relevant thread dump when in that state seems to be:

snapPuller-7-thread-1 (12)
java.util.concurrent.FutureTask$Sync@55ccfb48
sun.misc.Unsafe.park(Native Method)
java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt
(AbstractQueuedSynchronizer.java:811)
java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly
(AbstractQueuedSynchronizer.java:969)
java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly
(AbstractQueuedSynchronizer.java:1281)
java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:218)
java.util.concurrent.FutureTask.get(FutureTask.java:83)
org.apache.solr.handler.SnapPuller.doCommit(SnapPuller.java:655)
org.apache.solr.handler.SnapPuller.fetchLatestIndex(SnapPuller.java:466)
org.apache.solr.handler.ReplicationHandler.doFetch(ReplicationHandler.java:281)
org.apache.solr.handler.SnapPuller$1.run(SnapPuller.java:223)
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
java.util.concurrent.FutureTask$Sync.innerRunAndReset(FutureTask.java:317)
java.util.concurrent.FutureTask.runAndReset(FutureTask.java:150)
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$101
(ScheduledThreadPoolExecutor.java:98)
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.runPeriodic
(ScheduledThreadPoolExecutor.java:180)
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run
(ScheduledThreadPoolExecutor.java:204)
java.util.concurrent.ThreadPoolExecutor$Worker.runTask
(ThreadPoolExecutor.java:895)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
java.lang.Thread.run(Thread.java:662)

Tomcat containers are all the same; each of the slaves is running
entirely alone on its own container, separate machines. SOLR reported
versions:

Versions
solr-spec
4.2.0.2013.03.01.10.10.50
solr-impl
4.2-SNAPSHOT 1451604 - ensorn - 2013-03-01 10:10:50
lucene-spec
4.2-SNAPSHOT
lucene-impl
4.2-SNAPSHOT 1451604 - ensorn - 2013-03-01 10:02:53

Any help would be appreciated.  This is getting very frustrating.  To
make things worse, I have set up a new "slave" on my work PC (Mac),
and it replicates FLAWLESSLY on the same set up; only difference is
the slaves on the servers are on a SAN array (not sure if locking is
causing the heartburn?)  Any pointers would be great.  This is
obviously becoming a pain to work with, even with fairly infrequent
replications.

Thanks!

Neal Ensor
nen...@gmail.com