Nice proposal. I’d be interested in contributing as well. I’m about at my mentor limit with projects, but I’d be willing to contribute in other/similar ways.
-Taylor > On Jan 20, 2016, at 12:46 PM, Jean-Baptiste Onofré <j...@nanthrax.net> wrote: > > Great, I add you in the initial committer list then ;) > > I quickly discussed with James, we gonna create a section for additional > people as proposed by Sean. > > Thanks ! > Regards > JB > > On 01/20/2016 06:33 PM, Debo Dutta (dedutta) wrote: >> Hi JB >> >> Would love to join now. >> >> regards >> debo >> >> On 1/20/16, 9:31 AM, "Jean-Baptiste Onofré" <j...@nanthrax.net> wrote: >> >>> Hi Debo, >>> >>> Awesome: do you want to join now (in the initial committer list) and >>> once we are in the incubation ? >>> >>> Let me know, I can update the proposal. >>> >>> Regards >>> JB >>> >>> On 01/20/2016 06:23 PM, Debo Dutta (dedutta) wrote: >>>> +1 >>>> >>>> Proposal looks good. Also a small section on relationships with Apache >>>> Storm and Apache Samza would be great. >>>> >>>> I would like to sign up, to help/contribute. >>>> >>>> debo >>>> >>>> On 1/20/16, 8:55 AM, "Sean Busbey" <bus...@cloudera.com> wrote: >>>> >>>>> Great proposal. I like that your proposal includes a well presented >>>>> roadmap, but I don't see any goals that directly address building a >>>>> larger >>>>> community. Y'all have any ideas around outreach that will help with >>>>> adoption? >>>>> >>>>> As a start, I recommend y'all add a section to the proposal on the wiki >>>>> page for "Additional Interested Contributors" so that folks who want to >>>>> sign up to participate in the project can do so without requesting >>>>> additions to the initial committer list. >>>>> >>>>> On Wed, Jan 20, 2016 at 10:32 AM, James Malone < >>>>> jamesmal...@google.com.invalid> wrote: >>>>> >>>>>> Hello everyone, >>>>>> >>>>>> Attached to this message is a proposed new project - Apache Dataflow, >>>>>> a >>>>>> unified programming model for data processing and integration. >>>>>> >>>>>> The text of the proposal is included below. Additionally, the proposal >>>>>> is >>>>>> in draft form on the wiki where we will make any required changes: >>>>>> >>>>>> https://wiki.apache.org/incubator/DataflowProposal >>>>>> >>>>>> We look forward to your feedback and input. >>>>>> >>>>>> Best, >>>>>> >>>>>> James >>>>>> >>>>>> ---- >>>>>> >>>>>> = Apache Dataflow = >>>>>> >>>>>> == Abstract == >>>>>> >>>>>> Dataflow is an open source, unified model and set of language-specific >>>>>> SDKs >>>>>> for defining and executing data processing workflows, and also data >>>>>> ingestion and integration flows, supporting Enterprise Integration >>>>>> Patterns >>>>>> (EIPs) and Domain Specific Languages (DSLs). Dataflow pipelines >>>>>> simplify >>>>>> the mechanics of large-scale batch and streaming data processing and >>>>>> can >>>>>> run on a number of runtimes like Apache Flink, Apache Spark, and >>>>>> Google >>>>>> Cloud Dataflow (a cloud service). Dataflow also brings DSL in >>>>>> different >>>>>> languages, allowing users to easily implement their data integration >>>>>> processes. >>>>>> >>>>>> == Proposal == >>>>>> >>>>>> Dataflow is a simple, flexible, and powerful system for distributed >>>>>> data >>>>>> processing at any scale. Dataflow provides a unified programming >>>>>> model, >>>>>> a >>>>>> software development kit to define and construct data processing >>>>>> pipelines, >>>>>> and runners to execute Dataflow pipelines in several runtime engines, >>>>>> like >>>>>> Apache Spark, Apache Flink, or Google Cloud Dataflow. Dataflow can be >>>>>> used >>>>>> for a variety of streaming or batch data processing goals including >>>>>> ETL, >>>>>> stream analysis, and aggregate computation. The underlying programming >>>>>> model for Dataflow provides MapReduce-like parallelism, combined with >>>>>> support for powerful data windowing, and fine-grained correctness >>>>>> control. >>>>>> >>>>>> == Background == >>>>>> >>>>>> Dataflow started as a set of Google projects focused on making data >>>>>> processing easier, faster, and less costly. The Dataflow model is a >>>>>> successor to MapReduce, FlumeJava, and Millwheel inside Google and is >>>>>> focused on providing a unified solution for batch and stream >>>>>> processing. >>>>>> These projects on which Dataflow is based have been published in >>>>>> several >>>>>> papers made available to the public: >>>>>> >>>>>> * MapReduce - http://research.google.com/archive/mapreduce.html >>>>>> >>>>>> * Dataflow model - http://www.vldb.org/pvldb/vol8/p1792-Akidau.pdf >>>>>> >>>>>> * FlumeJava - http://notes.stephenholiday.com/FlumeJava.pdf >>>>>> >>>>>> * MillWheel - http://research.google.com/pubs/pub41378.html >>>>>> >>>>>> Dataflow was designed from the start to provide a portable programming >>>>>> layer. When you define a data processing pipeline with the Dataflow >>>>>> model, >>>>>> you are creating a job which is capable of being processed by any >>>>>> number of >>>>>> Dataflow processing engines. Several engines have been developed to >>>>>> run >>>>>> Dataflow pipelines in other open source runtimes, including a Dataflow >>>>>> runner for Apache Flink and Apache Spark. There is also a ³direct >>>>>> runner², >>>>>> for execution on the developer machine (mainly for dev/debug >>>>>> purposes). >>>>>> Another runner allows a Dataflow program to run on a managed service, >>>>>> Google Cloud Dataflow, in Google Cloud Platform. The Dataflow Java SDK >>>>>> is >>>>>> already available on GitHub, and independent from the Google Cloud >>>>>> Dataflow >>>>>> service. Another Python SDK is currently in active development. >>>>>> >>>>>> In this proposal, the Dataflow SDKs, model, and a set of runners will >>>>>> be >>>>>> submitted as an OSS project under the ASF. The runners which are a >>>>>> part >>>>>> of >>>>>> this proposal include those for Spark (from Cloudera), Flink (from >>>>>> data >>>>>> Artisans), and local development (from Google); the Google Cloud >>>>>> Dataflow >>>>>> service runner is not included in this proposal. Further references to >>>>>> Dataflow will refer to the Dataflow model, SDKs, and runners which >>>>>> are a >>>>>> part of this proposal (Apache Dataflow) only. The initial submission >>>>>> will >>>>>> contain the already-released Java SDK; Google intends to submit the >>>>>> Python >>>>>> SDK later in the incubation process. The Google Cloud Dataflow service >>>>>> will >>>>>> continue to be one of many runners for Dataflow, built on Google Cloud >>>>>> Platform, to run Dataflow pipelines. Necessarily, Cloud Dataflow will >>>>>> develop against the Apache project additions, updates, and changes. >>>>>> Google >>>>>> Cloud Dataflow will become one user of Apache Dataflow and will >>>>>> participate >>>>>> in the project openly and publicly. >>>>>> >>>>>> The Dataflow programming model has been designed with simplicity, >>>>>> scalability, and speed as key tenants. In the Dataflow model, you only >>>>>> need >>>>>> to think about four top-level concepts when constructing your data >>>>>> processing job: >>>>>> >>>>>> * Pipelines - The data processing job made of a series of computations >>>>>> including input, processing, and output >>>>>> >>>>>> * PCollections - Bounded (or unbounded) datasets which represent the >>>>>> input, >>>>>> intermediate and output data in pipelines >>>>>> >>>>>> * PTransforms - A data processing step in a pipeline in which one or >>>>>> more >>>>>> PCollections are an input and output >>>>>> >>>>>> * I/O Sources and Sinks - APIs for reading and writing data which are >>>>>> the >>>>>> roots and endpoints of the pipeline >>>>>> >>>>>> == Rationale == >>>>>> >>>>>> With Dataflow, Google intended to develop a framework which allowed >>>>>> developers to be maximally productive in defining the processing, and >>>>>> then >>>>>> be able to execute the program at various levels of >>>>>> latency/cost/completeness without re-architecting or re-writing it. >>>>>> This >>>>>> goal was informed by Google¹s past experience developing several >>>>>> models, >>>>>> frameworks, and tools useful for large-scale and distributed data >>>>>> processing. While Google has previously published papers describing >>>>>> some of >>>>>> its technologies, Google decided to take a different approach with >>>>>> Dataflow. Google open-sourced the SDK and model alongside >>>>>> commercialization >>>>>> of the idea and ahead of publishing papers on the topic. As a result, >>>>>> a >>>>>> number of open source runtimes exist for Dataflow, such as the Apache >>>>>> Flink >>>>>> and Apache Spark runners. >>>>>> >>>>>> We believe that submitting Dataflow as an Apache project will provide >>>>>> an >>>>>> immediate, worthwhile, and substantial contribution to the open source >>>>>> community. As an incubating project, we believe Dataflow will have a >>>>>> better >>>>>> opportunity to provide a meaningful contribution to OSS and also >>>>>> integrate >>>>>> with other Apache projects. >>>>>> >>>>>> In the long term, we believe Dataflow can be a powerful abstraction >>>>>> layer >>>>>> for data processing. By providing an abstraction layer for data >>>>>> pipelines >>>>>> and processing, data workflows can be increasingly portable, resilient >>>>>> to >>>>>> breaking changes in tooling, and compatible across many execution >>>>>> engines, >>>>>> runtimes, and open source projects. >>>>>> >>>>>> == Initial Goals == >>>>>> >>>>>> We are breaking our initial goals into immediate (< 2 months), >>>>>> short-term >>>>>> (2-4 months), and intermediate-term (> 4 months). >>>>>> >>>>>> Our immediate goals include the following: >>>>>> >>>>>> * Plan for reconciling the Dataflow Java SDK and various runners into >>>>>> one >>>>>> project >>>>>> >>>>>> * Plan for refactoring the existing Java SDK for better extensibility >>>>>> by >>>>>> SDK and runner writers >>>>>> >>>>>> * Validating all dependencies are ASL 2.0 or compatible >>>>>> >>>>>> * Understanding and adapting to the Apache development process >>>>>> >>>>>> Our short-term goals include: >>>>>> >>>>>> * Moving the newly-merged lists, and build utilities to Apache >>>>>> >>>>>> * Start refactoring codebase and move code to Apache Git repo >>>>>> >>>>>> * Continue development of new features, functions, and fixes in the >>>>>> Dataflow Java SDK, and Dataflow runners >>>>>> >>>>>> * Cleaning up the Dataflow SDK sources and crafting a roadmap and plan >>>>>> for >>>>>> how to include new major ideas, modules, and runtimes >>>>>> >>>>>> * Establishment of easy and clear build/test framework for Dataflow >>>>>> and >>>>>> associated runtimes; creation of testing, rollback, and validation >>>>>> policy >>>>>> >>>>>> * Analysis and design for work needed to make Dataflow a better data >>>>>> processing abstraction layer for multiple open source frameworks and >>>>>> environments >>>>>> >>>>>> Finally, we have a number of intermediate-term goals: >>>>>> >>>>>> * Roadmapping, planning, and execution of integrations with other OSS >>>>>> and >>>>>> non-OSS projects/products >>>>>> >>>>>> * Inclusion of additional SDK for Python, which is under active >>>>>> development >>>>>> >>>>>> == Current Status == >>>>>> >>>>>> === Meritocracy === >>>>>> >>>>>> Dataflow was initially developed based on ideas from many employees >>>>>> within >>>>>> Google. As an ASL OSS project on GitHub, the Dataflow SDK has received >>>>>> contributions from data Artisans, Cloudera Labs, and other individual >>>>>> developers. As a project under incubation, we are committed to >>>>>> expanding >>>>>> our effort to build an environment which supports a meritocracy. We >>>>>> are >>>>>> focused on engaging the community and other related projects for >>>>>> support >>>>>> and contributions. Moreover, we are committed to ensure contributors >>>>>> and >>>>>> committers to Dataflow come from a broad mix of organizations through >>>>>> a >>>>>> merit-based decision process during incubation. We believe strongly in >>>>>> the >>>>>> Dataflow model and are committed to growing an inclusive community of >>>>>> Dataflow contributors. >>>>>> >>>>>> === Community === >>>>>> >>>>>> The core of the Dataflow Java SDK has been developed by Google for use >>>>>> with >>>>>> Google Cloud Dataflow. Google has active community engagement in the >>>>>> SDK >>>>>> GitHub repository >>>>>> (https://github.com/GoogleCloudPlatform/DataflowJavaSDK >>>>>> ), >>>>>> on Stack Overflow ( >>>>>> http://stackoverflow.com/questions/tagged/google-cloud-dataflow) and >>>>>> has >>>>>> had contributions from a number of organizations and indivuduals. >>>>>> >>>>>> Everyday, Cloud Dataflow is actively used by a number of organizations >>>>>> and >>>>>> institutions for batch and stream processing of data. We believe >>>>>> acceptance >>>>>> will allow us to consolidate existing Dataflow-related work, grow the >>>>>> Dataflow community, and deepen connections between Dataflow and other >>>>>> open >>>>>> source projects. >>>>>> >>>>>> === Core Developers === >>>>>> >>>>>> The core developers for Dataflow and the Dataflow runners are: >>>>>> >>>>>> * Frances Perry >>>>>> >>>>>> * Tyler Akidau >>>>>> >>>>>> * Davor Bonaci >>>>>> >>>>>> * Luke Cwik >>>>>> >>>>>> * Ben Chambers >>>>>> >>>>>> * Kenn Knowles >>>>>> >>>>>> * Dan Halperin >>>>>> >>>>>> * Daniel Mills >>>>>> >>>>>> * Mark Shields >>>>>> >>>>>> * Craig Chambers >>>>>> >>>>>> * Maximilian Michels >>>>>> >>>>>> * Tom White >>>>>> >>>>>> * Josh Wills >>>>>> >>>>>> === Alignment === >>>>>> >>>>>> The Dataflow SDK can be used to create Dataflow pipelines which can be >>>>>> executed on Apache Spark or Apache Flink. Dataflow is also related to >>>>>> other >>>>>> Apache projects, such as Apache Crunch. We plan on expanding >>>>>> functionality >>>>>> for Dataflow runners, support for additional domain specific >>>>>> languages, >>>>>> and >>>>>> increased portability so Dataflow is a powerful abstraction layer for >>>>>> data >>>>>> processing. >>>>>> >>>>>> == Known Risks == >>>>>> >>>>>> === Orphaned Products === >>>>>> >>>>>> The Dataflow SDK is presently used by several organizations, from >>>>>> small >>>>>> startups to Fortune 100 companies, to construct production pipelines >>>>>> which >>>>>> are executed in Google Cloud Dataflow. Google has a long-term >>>>>> commitment to >>>>>> advance the Dataflow SDK; moreover, Dataflow is seeing increasing >>>>>> interest, >>>>>> development, and adoption from organizations outside of Google. >>>>>> >>>>>> === Inexperience with Open Source === >>>>>> >>>>>> Google believes strongly in open source and the exchange of >>>>>> information >>>>>> to >>>>>> advance new ideas and work. Examples of this commitment are active OSS >>>>>> projects such as Chromium (https://www.chromium.org) and Kubernetes ( >>>>>> http://kubernetes.io/). With Dataflow, we have tried to be >>>>>> increasingly >>>>>> open and forward-looking; we have published a paper in the VLDB >>>>>> conference >>>>>> describing the Dataflow model ( >>>>>> http://www.vldb.org/pvldb/vol8/p1792-Akidau.pdf) and were quick to >>>>>> release >>>>>> the Dataflow SDK as open source software with the launch of Cloud >>>>>> Dataflow. >>>>>> Our submission to the Apache Software Foundation is a logical >>>>>> extension >>>>>> of >>>>>> our commitment to open source software. >>>>>> >>>>>> === Homogeneous Developers === >>>>>> >>>>>> The majority of committers in this proposal belong to Google due to >>>>>> the >>>>>> fact that Dataflow has emerged from several internal Google projects. >>>>>> This >>>>>> proposal also includes committers outside of Google who are actively >>>>>> involved with other Apache projects, such as Hadoop, Flink, and Spark. >>>>>> We >>>>>> expect our entry into incubation will allow us to expand the number of >>>>>> individuals and organizations participating in Dataflow development. >>>>>> Additionally, separation of the Dataflow SDK from Google Cloud >>>>>> Dataflow >>>>>> allows us to focus on the open source SDK and model and do what is >>>>>> best >>>>>> for >>>>>> this project. >>>>>> >>>>>> === Reliance on Salaried Developers === >>>>>> >>>>>> The Dataflow SDK and Dataflow runners have been developed primarily by >>>>>> salaried developers supporting the Google Cloud Dataflow project. >>>>>> While >>>>>> the >>>>>> Dataflow SDK and Cloud Dataflow have been developed by different teams >>>>>> (and >>>>>> this proposal would reinforce that separation) we expect our initial >>>>>> set of >>>>>> developers will still primarily be salaried. Contribution has not been >>>>>> exclusively from salaried developers, however. For example, the >>>>>> contrib >>>>>> directory of the Dataflow SDK ( >>>>>> >>>>>> >>>>>> https://github.com/GoogleCloudPlatform/DataflowJavaSDK/tree/master/cont >>>>>> ri >>>>>> b >>>>>> ) >>>>>> contains items from free-time contributors. Moreover, seperate >>>>>> projects, >>>>>> such as ScalaFlow (https://github.com/darkjh/scalaflow) have been >>>>>> created >>>>>> around the Dataflow model and SDK. We expect our reliance on salaried >>>>>> developers will decrease over time during incubation. >>>>>> >>>>>> === Relationship with other Apache products === >>>>>> >>>>>> Dataflow directly interoperates with or utilizes several existing >>>>>> Apache >>>>>> projects. >>>>>> >>>>>> * Build >>>>>> >>>>>> ** Apache Maven >>>>>> >>>>>> * Data I/O, Libraries >>>>>> >>>>>> ** Apache Avro >>>>>> >>>>>> ** Apache Commons >>>>>> >>>>>> * Dataflow runners >>>>>> >>>>>> ** Apache Flink >>>>>> >>>>>> ** Apache Spark >>>>>> >>>>>> Dataflow when used in batch mode shares similarities with Apache >>>>>> Crunch; >>>>>> however, Dataflow is focused on a model, SDK, and abstraction layer >>>>>> beyond >>>>>> Spark and Hadoop (MapReduce.) One key goal of Dataflow is to provide >>>>>> an >>>>>> intermediate abstraction layer which can easily be implemented and >>>>>> utilized >>>>>> across several different processing frameworks. >>>>>> >>>>>> === An excessive fascination with the Apache brand === >>>>>> >>>>>> With this proposal we are not seeking attention or publicity. Rather, >>>>>> we >>>>>> firmly believe in the Dataflow model, SDK, and the ability to make >>>>>> Dataflow >>>>>> a powerful yet simple framework for data processing. While the >>>>>> Dataflow >>>>>> SDK >>>>>> and model have been open source, we believe putting code on GitHub can >>>>>> only >>>>>> go so far. We see the Apache community, processes, and mission as >>>>>> critical >>>>>> for ensuring the Dataflow SDK and model are truly community-driven, >>>>>> positively impactful, and innovative open source software. While >>>>>> Google >>>>>> has >>>>>> taken a number of steps to advance its various open source projects, >>>>>> we >>>>>> believe Dataflow is a great fit for the Apache Software Foundation due >>>>>> to >>>>>> its focus on data processing and its relationships to existing ASF >>>>>> projects. >>>>>> >>>>>> == Documentation == >>>>>> >>>>>> The following documentation is relevant to this proposal. Relevant >>>>>> portion >>>>>> of the documentation will be contributed to the Apache Dataflow >>>>>> project. >>>>>> >>>>>> * Dataflow website: https://cloud.google.com/dataflow >>>>>> >>>>>> * Dataflow programming model: >>>>>> https://cloud.google.com/dataflow/model/programming-model >>>>>> >>>>>> * Codebases >>>>>> >>>>>> ** Dataflow Java SDK: >>>>>> https://github.com/GoogleCloudPlatform/DataflowJavaSDK >>>>>> >>>>>> ** Flink Dataflow runner: >>>>>> https://github.com/dataArtisans/flink-dataflow >>>>>> >>>>>> ** Spark Dataflow runner: https://github.com/cloudera/spark-dataflow >>>>>> >>>>>> * Dataflow Java SDK issue tracker: >>>>>> https://github.com/GoogleCloudPlatform/DataflowJavaSDK/issues >>>>>> >>>>>> * google-cloud-dataflow tag on Stack Overflow: >>>>>> http://stackoverflow.com/questions/tagged/google-cloud-dataflow >>>>>> >>>>>> == Initial Source == >>>>>> >>>>>> The initial source for Dataflow which we will submit to the Apache >>>>>> Foundation will include several related projects which are currently >>>>>> hosted >>>>>> on the GitHub repositories: >>>>>> >>>>>> * Dataflow Java SDK ( >>>>>> https://github.com/GoogleCloudPlatform/DataflowJavaSDK) >>>>>> >>>>>> * Flink Dataflow runner >>>>>> (https://github.com/dataArtisans/flink-dataflow) >>>>>> >>>>>> * Spark Dataflow runner (https://github.com/cloudera/spark-dataflow) >>>>>> >>>>>> These projects have always been Apache 2.0 licensed. We intend to >>>>>> bundle >>>>>> all of these repositories since they are all complimentary and should >>>>>> be >>>>>> maintained in one project. Prior to our submission, we will combine >>>>>> all >>>>>> of >>>>>> these projects into a new git repository. >>>>>> >>>>>> == Source and Intellectual Property Submission Plan == >>>>>> >>>>>> The source for the Dataflow SDK and the three runners (Spark, Flink, >>>>>> Google >>>>>> Cloud Dataflow) are already licensed under an Apache 2 license. >>>>>> >>>>>> * Dataflow SDK - >>>>>> >>>>>> >>>>>> https://github.com/GoogleCloudPlatform/DataflowJavaSDK/blob/master/LICE >>>>>> NS >>>>>> E >>>>>> >>>>>> * Flink runner - >>>>>> https://github.com/dataArtisans/flink-dataflow/blob/master/LICENSE >>>>>> >>>>>> * Spark runner - >>>>>> https://github.com/cloudera/spark-dataflow/blob/master/LICENSE >>>>>> >>>>>> Contributors to the Dataflow SDK have also signed the Google >>>>>> Individual >>>>>> Contributor License Agreement ( >>>>>> https://cla.developers.google.com/about/google-individual) in order to >>>>>> contribute to the project. >>>>>> >>>>>> With respect to trademark rights, Google does not hold a trademark on >>>>>> the >>>>>> phrase ³Dataflow.² Based on feedback and guidance we receive during >>>>>> the >>>>>> incubation process, we are open to renaming the project if necessary >>>>>> for >>>>>> trademark or other concerns. >>>>>> >>>>>> == External Dependencies == >>>>>> >>>>>> All external dependencies are licensed under an Apache 2.0 or >>>>>> Apache-compatible license. As we grow the Dataflow community we will >>>>>> configure our build process to require and validate all contributions >>>>>> and >>>>>> dependencies are licensed under the Apache 2.0 license or are under an >>>>>> Apache-compatible license. >>>>>> >>>>>> == Required Resources == >>>>>> >>>>>> === Mailing Lists === >>>>>> >>>>>> We currently use a mix of mailing lists. We will migrate our existing >>>>>> mailing lists to the following: >>>>>> >>>>>> * d...@dataflow.incubator.apache.org >>>>>> >>>>>> * u...@dataflow.incubator.apache.org >>>>>> >>>>>> * priv...@dataflow.incubator.apache.org >>>>>> >>>>>> * comm...@dataflow.incubator.apache.org >>>>>> >>>>>> === Source Control === >>>>>> >>>>>> The Dataflow team currently uses Git and would like to continue to do >>>>>> so. >>>>>> We request a Git repository for Dataflow with mirroring to GitHub >>>>>> enabled. >>>>>> >>>>>> === Issue Tracking === >>>>>> >>>>>> We request the creation of an Apache-hosted JIRA. The Dataflow project >>>>>> is >>>>>> currently using both a public GitHub issue tracker and internal Google >>>>>> issue tracking. We will migrate and combine from these two sources to >>>>>> the >>>>>> Apache JIRA. >>>>>> >>>>>> == Initial Committers == >>>>>> >>>>>> * Aljoscha Krettek [aljos...@apache.org] >>>>>> >>>>>> * Amit Sela [amitsel...@gmail.com] >>>>>> >>>>>> * Ben Chambers [bchamb...@google.com] >>>>>> >>>>>> * Craig Chambers [chamb...@google.com] >>>>>> >>>>>> * Dan Halperin [dhalp...@google.com] >>>>>> >>>>>> * Davor Bonaci [da...@google.com] >>>>>> >>>>>> * Frances Perry [f...@google.com] >>>>>> >>>>>> * James Malone [jamesmal...@google.com] >>>>>> >>>>>> * Jean-Baptiste Onofré [jbono...@apache.org] >>>>>> >>>>>> * Josh Wills [jwi...@apache.org] >>>>>> >>>>>> * Kostas Tzoumas [kos...@data-artisans.com] >>>>>> >>>>>> * Kenneth Knowles [k...@google.com] >>>>>> >>>>>> * Luke Cwik [lc...@google.com] >>>>>> >>>>>> * Maximilian Michels [m...@apache.org] >>>>>> >>>>>> * Stephan Ewen [step...@data-artisans.com] >>>>>> >>>>>> * Tom White [t...@cloudera.com] >>>>>> >>>>>> * Tyler Akidau [taki...@google.com] >>>>>> >>>>>> == Affiliations == >>>>>> >>>>>> The initial committers are from six organizations. Google developed >>>>>> Dataflow and the Dataflow SDK, data Artisans developed the Flink >>>>>> runner, >>>>>> and Cloudera (Labs) developed the Spark runner. >>>>>> >>>>>> * Cloudera >>>>>> >>>>>> ** Tom White >>>>>> >>>>>> * Data Artisans >>>>>> >>>>>> ** Aljoscha Krettek >>>>>> >>>>>> ** Kostas Tzoumas >>>>>> >>>>>> ** Maximilian Michels >>>>>> >>>>>> ** Stephan Ewen >>>>>> >>>>>> * Google >>>>>> >>>>>> ** Ben Chambers >>>>>> >>>>>> ** Dan Halperin >>>>>> >>>>>> ** Davor Bonaci >>>>>> >>>>>> ** Frances Perry >>>>>> >>>>>> ** James Malone >>>>>> >>>>>> ** Kenneth Knowles >>>>>> >>>>>> ** Luke Cwik >>>>>> >>>>>> ** Tyler Akidau >>>>>> >>>>>> * PayPal >>>>>> >>>>>> ** Amit Sela >>>>>> >>>>>> * Slack >>>>>> >>>>>> ** Josh Wills >>>>>> >>>>>> * Talend >>>>>> >>>>>> ** Jean-Baptiste Onofré >>>>>> >>>>>> == Sponsors == >>>>>> >>>>>> === Champion === >>>>>> >>>>>> * Jean-Baptiste Onofre [jbono...@apache.org] >>>>>> >>>>>> === Nominated Mentors === >>>>>> >>>>>> * Jim Jagielski [j...@apache.org] >>>>>> >>>>>> * Venkatesh Seetharam [venkat...@apache.org] >>>>>> >>>>>> * Bertrand Delacretaz [bdelacre...@apache.org] >>>>>> >>>>>> * Ted Dunning [tdunn...@apache.org] >>>>>> >>>>>> === Sponsoring Entity === >>>>>> >>>>>> The Apache Incubator >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Sean >>>> >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org >>>> For additional commands, e-mail: general-h...@incubator.apache.org >>>> >>> >>> -- >>> Jean-Baptiste Onofré >>> jbono...@apache.org >>> http://blog.nanthrax.net >>> Talend - http://www.talend.com >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org >>> For additional commands, e-mail: general-h...@incubator.apache.org >>> >> >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org >> For additional commands, e-mail: general-h...@incubator.apache.org >> > > -- > Jean-Baptiste Onofré > jbono...@apache.org > http://blog.nanthrax.net > Talend - http://www.talend.com > > --------------------------------------------------------------------- > To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org > For additional commands, e-mail: general-h...@incubator.apache.org >
signature.asc
Description: Message signed with OpenPGP using GPGMail