I'm not following what a "simple operation replication framework" is and how it applies to the Redis API. If you replicate operations, you still need to update the data at some point, causing a synchronous replication event so as to provide HA. What is, in more detail, a "simple operation replication framework"?
Regards, Wes Williams Sent from mobile phone > On Feb 27, 2017, at 2:07 PM, Bruce Schuchardt <bschucha...@pivotal.io> wrote: > > What I like about using the existing delta-prop mechanism is it will also > extend to client subscriptions & WAN. It would take a lot of work to > propagate Redis commands through those paths. > >> Le 2/27/2017 à 10:35 AM, Hitesh Khamesra a écrit : >> The simplicity of Redis API making this problem (delta update)most apparent. >> But, I would imagine many Geode apps has a similar use case. >> >> -Hitesh >> >> From: Michael Stolz <mst...@pivotal.io> >> To: dev@geode.apache.org >> Sent: Monday, February 27, 2017 9:06 AM >> Subject: Re: [DISCUSS] changes to Redis implementation >> It does seem like the operations will often be much smaller than the data >> they are operating on. >> It is almost the classic "move the code to the data" pattern. >> >> -- >> Mike Stolz >> Principal Engineer, GemFire Product Manager >> Mobile: +1-631-835-4771 >> >> On Mon, Feb 27, 2017 at 10:51 AM, Udo Kohlmeyer <ukohlme...@pivotal.io> >> wrote: >> >>> Delta could provide us a mechanism to replicate only what is required. >>> >>> I wonder if we could not create a simple operation replication framework. >>> Rather than writing a potential large amounts of code for delta, we >>> replicate only the operation. >>> >>> >>>> On 2/27/17 07:18, Wes Williams wrote: >>>> >>>> Replicating a whole collection because of 1 change does not really make >>>>> too much sense.<< >>>> I agree but won't delta replication prevent sending the entire collection >>>> across the wire? >>>> >>>> *Wes Williams | Pivotal Advisory **Data Engineer* >>>> 781.606.0325 >>>> http://pivotal.io/big-data/pivotal-gemfire >>>> >>>> On Mon, Feb 27, 2017 at 10:08 AM, Udo Kohlmeyer <ukohlme...@pivotal.io> >>>> wrote: >>>> >>>> I've quickly gone through the changes for the pull request. >>>>> The most significant change of this pull request is that the collections >>>>> that initially were regions are single collections (not distributed). >>>>> That >>>>> said, this is something that we've been discussing. The one thing that I >>>>> wonder about is, what will the performance look like when the collections >>>>> become really large? Replicating a whole collection because of 1 change >>>>> does not really make too much sense. >>>>> >>>>> Maybe this implementation becomes the catalyst for future improvements. >>>>> >>>>> --Udo >>>>> >>>>> >>>>> >>>>> On 2/24/17 15:25, Bruce Schuchardt wrote: >>>>> >>>>> Gregory Green has posted a pull request that warrants discussion. It >>>>>> improves performance for Sets and Hashes by altering the storage format >>>>>> for >>>>>> these collections. As such it will not permit a rolling upgrade, though >>>>>> the Redis adapter is labelled "experimental" so maybe that's okay. >>>>>> >>>>>> https://github.com/apache/geode/pull/404 >>>>>> >>>>>> The PR also fixes GEODE-2469, inability to handle hash keys having >>>>>> colons. >>>>>> >>>>>> There was some discussion about altering the storage format that was >>>>>> initiated by Hitesh. Personally I think Gregory's changes are better >>>>>> than >>>>>> the current implementation and we should accept them, though I haven't >>>>>> gone >>>>>> through the code changes extensively. >>>>>> >>>>>> >>>>>> >> >> >