Hi list,

We have deployed an experimental Solr 1.4 cluster (a master/slave
setup, with automatic promotion of the slave as a master in case of
failure) on drupal.org, to manage our medium size index (3GB, about
400K documents).

One of the problem we are facing is that there seems to be no sanity
checks before downloading files. Take the following scenario:

 - initial situation: s1 is master, s2 is slave
 - s1 fails, the virtual IP falls back to s2
 - some updates happen on s2
 - suppose now that s1 gets back online, s2 tries to replicate from
s1, but after replicating all the files (3GB), the commit fails
because the local index has been locally updated, the replication
fails, but the process restarts at the next poll (redownload all the
index files, fails again...) and so on

We are considering configuring each server to replicate from the
virtual IP, which should solve that issue for us, but couldn't the
slave do some sanity checks before trying to download all the files
from the master?

Thanks in advance for any help you could provide,

Damien Tournoud

Reply via email to