Re: solr replication vs. rsync

2015-01-25 Thread Erick Erickson
t; > > because they had talked about "replicating" the data. > > > > > > > > But we are not Pinterest, and do not expect to be taking in changes > one > > > > post at a time (at least the engineers don't - just wait until its > us

Re: solr replication vs. rsync

2015-01-25 Thread Dan Davis
Pinterest, and do not expect to be taking in changes one > > > post at a time (at least the engineers don't - just wait until its used > > for > > > a Crud app that wants full-text search on a description field!). > > Still, > > > rsync can be very, very

Re: solr replication vs. rsync

2015-01-25 Thread Dan Davis
one of them were using Solr replication, > > > because they had talked about "replicating" the data. > > > > > > But we are not Pinterest, and do not expect to be taking in changes one > > > post at a time (at least the engineers don't - just wait unti

Re: solr replication vs. rsync

2015-01-25 Thread Erick Erickson
ot expect to be taking in changes one > > post at a time (at least the engineers don't - just wait until its used > for > > a Crud app that wants full-text search on a description field!). > Still, > > rsync can be very, very fast with the right options (-W for gigabit &

Re: solr replication vs. rsync

2015-01-25 Thread Shawn Heisey
est, and do not expect to be taking in changes one > post at a time (at least the engineers don't - just wait until its used for > a Crud app that wants full-text search on a description field!).Still, > rsync can be very, very fast with the right options (-W for gigabit > ethern

solr replication vs. rsync

2015-01-24 Thread Dan Davis
(at least the engineers don't - just wait until its used for a Crud app that wants full-text search on a description field!). Still, rsync can be very, very fast with the right options (-W for gigabit ethernet, and maybe -S for sparse files). I've clocked it at 48 MB/s over GigE prev

Re: Solr 4.2.1 + Distribution scripts (rsync) Issue

2013-06-05 Thread Sandeep Mestry
Hi Hoss, Thanks for your reply, Please find answers to your questions below. *Well, for starters -- have you considered at least looking into using the java based Replicationhandler instead of the rsync scripts?* - There was an attempt to to implement java based replication but it was very slow

Re: Solr 4.2.1 + Distribution scripts (rsync) Issue

2013-06-04 Thread Chris Hostetter
: However, we haven't yet implemented SolrCloud and still relying on : distribution scripts - rsync, indexpuller mechanism. Well, for starters -- have you considered at least looking into using hte java based Replicationhandler instead of the rsync scripts? Script based replication ha

Solr 4.2.1 + Distribution scripts (rsync) Issue

2013-06-04 Thread Sandeep Mestry
ripts - rsync, indexpuller mechanism. *Issue:* I see that the indexes are getting created on indexer boxes, snapshots being created and then pulled across to search boxes. The snapshots are getting installed on search boxes as well. There are no errors in the scripts logs and this process works

Re: OS memory Pollution on rsync?

2012-06-03 Thread Otis Gospodnetic
Sujatha, A few thoughts: * If you are still using rsync replication, you may be using an old version of Solr.  Consider upgrading (it's also more efficient with memory). * Yes, rsync will "pollute" OS cache, but is it really pollution if it makes the OS cache the index that is a

OS memory Pollution on rsync?

2012-06-03 Thread Sujatha Arun
Hello, When we migrated the solr webapps to a new server in production,We did an rsync of all the indexes in our old server . On checking with Jconsole ,the OS free Ram was already filled to twice the Index size .There is no other service in this server. The rsync was done twice. Even though

Re: Master Slave data distribution | rsync fail issue

2009-07-02 Thread Bill Au
rediffmail.com> wrote: > > > >> > >> Hi, > >> > >> I am facing an issue while performing snapshot pulling thru Snappuller > >> script from slave server : > >> We have the setup of multicores on Master Solr and Slave Solr servers

Re: Master Slave data distribution | rsync fail issue

2009-07-02 Thread Vicky_Dev
ushar_kapoor...@rediffmail.com> wrote: > >> >> Hi, >> >> I am facing an issue while performing snapshot pulling thru Snappuller >> script from slave server : >> We have the setup of multicores on Master Solr and Slave Solr servers. >> Scenario , 2 cores

Re: Master Slave data distribution | rsync fail issue

2009-05-06 Thread Bill Au
pshot pulling thru Snappuller > script from slave server : > We have the setup of multicores on Master Solr and Slave Solr servers. > Scenario , 2 cores are set : > i) CORE_WWW.ABCD.COM > ii) CORE_WWW.XYZ.COM > > rsync-enable and rsync-start script run from CORE_WWW.ABCD.COM on

Master Slave data distribution | rsync fail issue

2009-05-05 Thread tushar kapoor
Hi, I am facing an issue while performing snapshot pulling thru Snappuller script from slave server : We have the setup of multicores on Master Solr and Slave Solr servers. Scenario , 2 cores are set : i) CORE_WWW.ABCD.COM ii) CORE_WWW.XYZ.COM rsync-enable and rsync-start script run

Re: rsync snappuller slowdown Qtime

2009-03-13 Thread Yonik Seeley
On Fri, Mar 13, 2009 at 10:33 AM, sunnyfr wrote: > And it was obvious how during snappuller on a slave server, the query time > was a lot longer than the rest of the time. Did the CPU utilization drop? It could be writing of the new files being pulled forcing parts of the current index files out

rsync snappuller slowdown Qtime

2009-03-13 Thread sunnyfr
round 200msec with snappuller 3-6sec .. Do you have any idea? Thanks a lot, -- View this message in context: http://www.nabble.com/rsync-snappuller-slowdown-Qtime-tp22497625p22497625.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: rsync

2009-02-26 Thread Otis Gospodnetic
- Nutch - Original Message > From: revas > To: solr-user@lucene.apache.org > Sent: Friday, February 27, 2009 12:07:50 AM > Subject: rsync > > When we rsync from master to slave ,since this refers to the same index that > is on the master ,if master server goes down

Re: snappuller / rsync

2008-09-22 Thread Bill Au
p 22, 2008 at 6:49 AM, sunnyfr <[EMAIL PROTECTED]> wrote: > > > >> > >> Yes but even if I run it I've no snapshot created, I don't get how I can > >> fixe > >> it. > >> > >> > >> > >> > >> Bill Au w

Re: snappuller / rsync

2008-09-22 Thread sunnyfr
snapshot created, I don't get how I can >> fixe >> it. >> >> >> >> >> Bill Au wrote: >> > >> > You only need to run the rsync daemon on the master. >> > >> > Bill >> > >> > On Wed, Sep

Re: snappuller / rsync

2008-09-22 Thread Bill Au
> > Bill Au wrote: > > > > You only need to run the rsync daemon on the master. > > > > Bill > > > > On Wed, Sep 17, 2008 at 10:54 AM, sunnyfr <[EMAIL PROTECTED]> wrote: > > > >> > >> Hi Raghu, > >> > >> Thanks

Re: snappuller / rsync

2008-09-22 Thread sunnyfr
Yes but even if I run it I've no snapshot created, I don't get how I can fixe it. Bill Au wrote: > > You only need to run the rsync daemon on the master. > > Bill > > On Wed, Sep 17, 2008 at 10:54 AM, sunnyfr <[EMAIL PROTECTED]> wrote: > >&g

Re: snappuller / rsync

2008-09-17 Thread Bill Au
You only need to run the rsync daemon on the master. Bill On Wed, Sep 17, 2008 at 10:54 AM, sunnyfr <[EMAIL PROTECTED]> wrote: > > Hi Raghu, > > Thanks it's clear now; > > > Kashyap, Raghu wrote: > > > > Hi, > > > > Rsyncd is the rsync(

RE: snappuller / rsync

2008-09-17 Thread sunnyfr
Hi Raghu, Thanks it's clear now; Kashyap, Raghu wrote: > > Hi, > > Rsyncd is the rsync(http://samba.anu.edu.au/rsync/) daemon. You need > to make sure that Rsynchd is running on both master & the slave > machines. You use snapshooter on the master server to

RE: snappuller / rsync

2008-09-17 Thread Kashyap, Raghu
Hi, Rsyncd is the rsync(http://samba.anu.edu.au/rsync/) daemon. You need to make sure that Rsynchd is running on both master & the slave machines. You use snapshooter on the master server to create the snapshot & run snappuller on the slave machines to pull those snapshots from maste

snappuller / rsync

2008-09-17 Thread sunnyfr
snappuller---rsync-tp19532941p19532941.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: force rsync update on a read-only index

2008-05-08 Thread Brian Whitman
utch - Original Message From: Brian Whitman <[EMAIL PROTECTED]> To: solr-user@lucene.apache.org Sent: Wednesday, May 7, 2008 11:25:22 AM Subject: force rsync update on a read-only index We have a few slave solr servers that are just hardlinked-rsynced copies of a master server

Re: force rsync update on a read-only index

2008-05-08 Thread Shalin Shekhar Mangar
e.apache.org > > Sent: Wednesday, May 7, 2008 11:25:22 AM > > Subject: force rsync update on a read-only index > > > > We have a few slave solr servers that are just hardlinked-rsynced > > copies of a master server. > > > > When we do the rsync the changes don

Re: force rsync update on a read-only index

2008-05-08 Thread Otis Gospodnetic
[EMAIL PROTECTED]> > To: solr-user@lucene.apache.org > Sent: Wednesday, May 7, 2008 11:25:22 AM > Subject: force rsync update on a read-only index > > We have a few slave solr servers that are just hardlinked-rsynced > copies of a master server. > > When we do the rsync the

force rsync update on a read-only index

2008-05-07 Thread Brian Whitman
We have a few slave solr servers that are just hardlinked-rsynced copies of a master server. When we do the rsync the changes don't show up immediately. The snap* scripts call a commit on the slave servers -- but since these are readonly servers we've disabled /update in the solr

RE: snappuller rsync parameter error? - "solr" hardcoded

2007-11-14 Thread Chris Hostetter
: Be careful. 'rsync' has different meanings for 'directory' v.s. : 'directory/'. I ran afoul of this. eliminating some of the confusion about how rsync is one of the reasons the rsyncd-start and snappuller scripts exist, and why a seperate rsyncd port is

RE: snappuller rsync parameter error? - "solr" hardcoded

2007-11-14 Thread Norskog, Lance
Be careful. 'rsync' has different meanings for 'directory' v.s. 'directory/'. I ran afoul of this. -Original Message- From: Walter Underwood [mailto:[EMAIL PROTECTED] Sent: Wednesday, November 14, 2007 8:49 AM To: solr-user@lucene.apache.org Subject: Re: sn

Re: snappuller rsync parameter error? - "solr" hardcoded

2007-11-14 Thread Chris Hostetter
: I'm not an rsync expert, but I beleive that /solr/ is a : virtual directory defined in the rsyncd config. It is mapped : to the real directory. correct .. the real directory comes from the scripts.conf : : wunder : : On 11/14/07 8:43 AM, "Jae Joo" <[EMAIL PROTECTED]>

Re: snappuller rsync parameter error? - "solr" hardcoded

2007-11-14 Thread Walter Underwood
I'm not an rsync expert, but I beleive that /solr/ is a virtual directory defined in the rsyncd config. It is mapped to the real directory. wunder On 11/14/07 8:43 AM, "Jae Joo" <[EMAIL PROTECTED]> wrote: > In the snappuller, the "solr" is hardcoded.

snappuller rsync parameter error? - "solr" hardcoded

2007-11-14 Thread Jae Joo
In the snappuller, the "solr" is hardcoded. Should it be "${master_data_dir}? # rsync over files that have changed rsync -Wa${verbose}${compress} --delete ${sizeonly} \ ${stats} rsync://${master_host}:${rsyncd_port}/solr/${name}/ ${data_dir}/${name}-wip Thanks, Jae

Re: small rsync index question

2007-09-28 Thread Chris Hostetter
: To completely remove the window of inconsistency, comment out the : post-commit hook in solrconfig.xml that takes a snapshot, then send a : commit to get a new snapshot and rsync from that. i think yonik ment "UN-comment" the postCommit hook in the example solrconfig.xml. -Hoss

Re: small rsync index question

2007-09-28 Thread Yonik Seeley
ery active, is > constantly getting new docs and searched on. What can I do to > preserve the index state while syncing? To greatly reduce this window, take a "snapshot" of the index (do it like snapshooter, or actually use that script) and rsync from that snapshot instead. To complet

Re: small rsync index question

2007-09-28 Thread Brian Whitman
Sep 28, 2007, at 5:41 PM, Yonik Seeley wrote: It should... are there any errors in the logs? do you see the commit in the logs? Check the stats page to see info about when the current searcher was last opened too. ugh, nevermind.. was committing the wrong solr index... but Thanks yonik fo

Re: small rsync index question

2007-09-28 Thread Yonik Seeley
On 9/28/07, Brian Whitman <[EMAIL PROTECTED]> wrote: > For some reason sending a > is not refreshing the index It should... are there any errors in the logs? do you see the commit in the logs? Check the stats page to see info about when the current searcher was last opened too. -Yonik

small rsync index question

2007-09-28 Thread Brian Whitman
I'm not using snap* scripts but i quickly need to sync up two indexes on two machines. I am rsyncing the data dirs from A to B, which work fine. But how can I see the new index on B? For some reason sending a is not refreshing the index, and I have to restart resin to see it. Is there some

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-21 Thread Bill Au
You are welcome. Bill On 9/21/07, Yu-Hui Jin <[EMAIL PROTECTED]> wrote: > > Bill, > > Thanks for the explanation. That helps my understanding on rsync and the > replication in general. > > > regards, > > -Hui > > On 9/20/07, Bill Au <[EMAIL PROTE

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-20 Thread Yu-Hui Jin
Bill, Thanks for the explanation. That helps my understanding on rsync and the replication in general. regards, -Hui On 9/20/07, Bill Au <[EMAIL PROTECTED]> wrote: > > The "solr" that you are referring to in your third question in the > name of the rsync area whic

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-20 Thread Bill Au
The "solr" that you are referring to in your third question in the name of the rsync area which is map to the solr data directory. This is defined in the rsyncd configuration file which is generated on the fly as Chris has pointed out. Take a look at rsyncd-start. snappuller rsync

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-20 Thread Chris Hostetter
: So just to help my knowledge, where does this virtual setting of this "solr" : string happen? Should it be in some config file or sth? rsyncd-start creates an rsync config file on the fly ... much of it is constants, but it fills in the rsync port using a variable from your config. -Hoss

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-20 Thread Yu-Hui Jin
r cases they usually use ${solr_root} to get to specific > dirs. > : ) I pasted this line again below: > > sorry ... i didn't realize you were talking about the script, i thought > you were talking aboutthe docs. > > : I saw in snappuller line 226 (solr 1.2): > : > :

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-20 Thread Chris Hostetter
didn't realize you were talking about the script, i thought you were talking aboutthe docs. : I saw in snappuller line 226 (solr 1.2): : : ${stats} rsync://${master_host}:${rsyncd_port}/solr/${name}/ : ${data_dir}/${name}-wip : : Is the above "solr" a hard-coded solr home name? If s

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-20 Thread Yu-Hui Jin
${solr_root} to get to specific dirs. ) I pasted this line again below: I saw in snappuller line 226 (solr 1.2): ${stats} rsync://${master_host}:${rsyncd_port}/solr/${name}/ ${data_dir}/${name}-wip Is the above "solr" a hard-coded solr home name? If so, it's not desirable sin

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-20 Thread Chris Hostetter
: 1) config different port for each solr home dir (since they run on the same : host); you mean a differnet rsync port right? ... yes the scripts as distributed assume that each rsync daemon will be dedicated to a single solr "instance" .. the idea beaing that even if you have 12 Sol

Re: rsync start and enable for multiple solr instances within one tomcat

2007-09-20 Thread Yu-Hui Jin
Ok, I should correct myself. For #1, I think we need to 1) config different port for each solr home dir (since they run on the same host); 2) run rsync-start script under each of the solr home's bin dir. (btw, just to make clear, we should run rsync-start after rsync-enable that I under

rsync start and enable for multiple solr instances within one tomcat

2007-09-19 Thread Yu-Hui Jin
Hi, there, So we are using the Tomcat's JNDI method to set up multiple solr instances within a tomcat server. Each instance has a solr home directory. Now we want to set up collection distribution for all these solr home indexes. My understanding is: 1. we only need to run rsync-start onc

Re: Rsync

2006-03-29 Thread Yonik Seeley
reloading would > make this more difficult. rsync will only copy over the changed index files, not the whole index each time. -Yonik

Re: Rsync

2006-03-29 Thread jason rutherglen
: Bill Au <[EMAIL PROTECTED]> To: solr-user@lucene.apache.org; jason rutherglen <[EMAIL PROTECTED]> Sent: Wednesday, March 29, 2006 6:07:35 AM Subject: Re: Rsync The segments file will change when new segments are created. But what I really meant before was that the file deletable also

Re: Rsync

2006-03-29 Thread Bill Au
slave > > Lucene can certainly record a list of all new segment files added. I > think the tricky part > is to ensure that a consistent copy of the index is being distributed. > > Bill > > > On 3/27/06, jason rutherglen <[EMAIL PROTECTED]> wrote: > > > >

Re: Rsync

2006-03-28 Thread Bill Au
ble to avoid using rsync and record a > list of all new segment files added (from within Lucene), and simply use > HTTP to sync down the newest ones? Perhaps only using rsync after an > optimize? Seems like if I understand Lucene correctly only new files are > created? > > >

Rsync

2006-03-27 Thread jason rutherglen
I was thinking, would it not be possible to avoid using rsync and record a list of all new segment files added (from within Lucene), and simply use HTTP to sync down the newest ones? Perhaps only using rsync after an optimize? Seems like if I understand Lucene correctly only new files are