Confirmed, and please update my affiliation to 'Qihoo 360'. Thanks. 2017-01-14 16:08 GMT+08:00 梁德澎 <liangdep...@gmail.com>: > Hi, > > I’ve been working on the MXNet-ScalaPkg for a while with Yizhi Liu > (@javelinjs). > Please sign me up as a committer of MxNet. > > GitHub ID: Ldpe2G > Email: liangdep...@gmail.com > Affiliations: Sun Yat-sen University > > 2017-01-14 13:49 GMT+08:00 Henri Yandell <bay...@apache.org>: > >> Thanks for all the feedback and interested parties :) >> >> My aim is to propose a vote on Monday, unless someone raises an issue >> before then. >> >> Hen >> >> On Thu, Jan 5, 2017 at 9:12 PM, Henri Yandell <bay...@apache.org> wrote: >> >> > Hello Incubator, >> > >> > I'd like to propose a new incubator Apache MXNet podling. >> > >> > The existing MXNet project (http://mxnet.io - 1.5 years old, 15 >> > committers, 200 contributors) is very interested in joining Apache. MXNet >> > is an open-source deep learning framework that allows you to define, >> train, >> > and deploy deep neural networks on a wide array of devices, from cloud >> > infrastructure to mobile devices. >> > >> > The wiki proposal page is located here: >> > >> > https://wiki.apache.org/incubator/MXNetProposal >> > >> > I've included the text below in case anyone wants to focus on parts of it >> > in a reply. >> > >> > Looking forward to your thoughts, and for lots of interested Apache >> > members to volunteer to mentor the project in addition to Sebastian and >> > myself. >> > >> > Currently the list of committers is based on the current active coders, >> so >> > we're also very interested in hearing from anyone else who is interested >> in >> > working on the project, be they current or future contributor! >> > >> > Thanks, >> > >> > Hen >> > On behalf of the MXNet project >> > >> > --------- >> > >> > = MXNet: Apache Incubator Proposal = >> > >> > == Abstract == >> > >> > MXNet is a Flexible and Efficient Library for Deep Learning >> > >> > == Proposal == >> > >> > MXNet is an open-source deep learning framework that allows you to >> define, >> > train, and deploy deep neural networks on a wide array of devices, from >> > cloud infrastructure to mobile devices. It is highly scalable, allowing >> for >> > fast model training, and supports a flexible programming model and >> multiple >> > languages. MXNet allows you to mix symbolic and imperative programming >> > flavors to maximize both efficiency and productivity. MXNet is built on a >> > dynamic dependency scheduler that automatically parallelizes both >> symbolic >> > and imperative operations on the fly. A graph optimization layer on top >> of >> > that makes symbolic execution fast and memory efficient. The MXNet >> library >> > is portable and lightweight, and it scales to multiple GPUs and multiple >> > machines. >> > >> > == Background == >> > >> > Deep learning is a subset of Machine learning and refers to a class of >> > algorithms that use a hierarchical approach with non-linearities to >> > discover and learn representations within data. Deep Learning has >> recently >> > become very popular due to its applicability and advancement of domains >> > such as Computer Vision, Speech Recognition, Natural Language >> Understanding >> > and Recommender Systems. With pervasive and cost effective cloud >> computing, >> > large labeled datasets and continued algorithmic innovation, Deep >> Learning >> > has become the one of the most popular classes of algorithms for machine >> > learning practitioners in recent years. >> > >> > == Rational == >> > >> > The adoption of deep learning is quickly expanding from initial deep >> > domain experts rooted in academia to data scientists and developers >> working >> > to deploy intelligent services and products. Deep learning however has >> many >> > challenges. These include model training time (which can take days to >> > weeks), programmability (not everyone writes Python or C++ and like >> > symbolic programming) and balancing production readiness (support for >> > things like failover) with development flexibility (ability to program >> > different ways, support for new operators and model types) and speed of >> > execution (fast and scalable model training). Other frameworks excel on >> > some but not all of these aspects. >> > >> > >> > == Initial Goals == >> > >> > MXNet is a fairly established project on GitHub with its first code >> > contribution in April 2015 and roughly 200 contributors. It is used by >> > several large companies and some of the top research institutions on the >> > planet. Initial goals would be the following: >> > >> > 1. Move the existing codebase(s) to Apache >> > 1. Integrate with the Apache development process/sign CLAs >> > 1. Ensure all dependencies are compliant with Apache License version 2.0 >> > 1. Incremental development and releases per Apache guidelines >> > 1. Establish engineering discipline and a predictable release cadence of >> > high quality releases >> > 1. Expand the community beyond the current base of expert level users >> > 1. Improve usability and the overall developer/user experience >> > 1. Add additional functionality to address newer problem types and >> > algorithms >> > >> > >> > == Current Status == >> > >> > === Meritocracy === >> > >> > The MXNet project already operates on meritocratic principles. Today, >> > MXNet has developers worldwide and has accepted multiple major patches >> from >> > a diverse set of contributors within both industry and academia. We would >> > like to follow ASF meritocratic principles to encourage more developers >> to >> > contribute in this project. We know that only active and committed >> > developers from a diverse set of backgrounds can make MXNet a successful >> > project. We are also improving the documentation and code to help new >> > developers get started quickly. >> > >> > === Community === >> > >> > Acceptance into the Apache foundation would bolster the growing user and >> > developer community around MXNet. That community includes around 200 >> > contributors from academia and industry. The core developers of our >> project >> > are listed in our contributors below and are also represented by logos on >> > the mxnet.io site including Amazon, Baidu, Carnegie Mellon University, >> > Turi, Intel, NYU, Nvidia, MIT, Microsoft, TuSimple, University of >> Alberta, >> > University of Washington and Wolfram. >> > >> > === Core Developers === >> > >> > (with GitHub logins) >> > >> > * Tianqi Chen (@tqchen) >> > * Mu Li (@mli) >> > * Junyuan Xie (@piiswrong) >> > * Bing Xu (@antinucleon) >> > * Chiyuan Zhang (@pluskid) >> > * Minjie Wang (@jermainewang) >> > * Naiyan Wang (@winstywang) >> > * Yizhi Liu (@javelinjs) >> > * Tong He (@hetong007) >> > * Qiang Kou (@thirdwing) >> > * Xingjian Shi (@sxjscience) >> > >> > === Alignment === >> > >> > ASF is already the home of many distributed platforms, e.g., Hadoop, >> Spark >> > and Mahout, each of which targets a different application domain. MXNet, >> > being a distributed platform for large-scale deep learning, focuses on >> > another important domain for which there still lacks a scalable, >> > programmable, flexible and super fast open-source platform. The recent >> > success of deep learning models especially for vision and speech >> > recognition tasks has generated interests in both applying existing deep >> > learning models and in developing new ones. Thus, an open-source platform >> > for deep learning backed by some of the top industry and academic players >> > will be able to attract a large community of users and developers. MXNet >> is >> > a complex system needing many iterations of design, implementation and >> > testing. Apache's collaboration framework which encourages active >> > contribution from developers will inevitably help improve the quality of >> > the system, as shown in the success of Hadoop, Spark, etc. Equally >> > important is the community of users which helps identify real-life >> > applications of deep learning, and helps to evaluate the system's >> > performance and ease-of-use. We hope to leverage ASF for coordinating and >> > promoting both communities, and in return benefit the communities with >> > another useful tool. >> > >> > == Known Risks == >> > >> > === Orphaned products === >> > >> > Given the current level of investment in MXNet and the stakeholders using >> > it - the risk of the project being abandoned is minimal. Amazon, for >> > example, is in active development to use MXNet in many of its services >> and >> > many large corporations use it in their production applications. >> > >> > === Inexperience with Open Source === >> > >> > MXNet has existed as a healthy open source project for more than a year. >> > During that time, the project has attracted 200+ contributors. >> > >> > === Homogenous Developers === >> > >> > The initial list of committers and contributors includes developers from >> > several institutions and industry participants (see above). >> > >> > === Reliance on Salaried Developers === >> > >> > Like most open source projects, MXNet receives a substantial support from >> > salaried developers. A large fraction of MXNet development is supported >> by >> > graduate students at various universities in the course of research >> degrees >> > - this is more a “volunteer” relationship, since in most cases students >> > contribute vastly more than is necessary to immediately support research. >> > In addition, those working from within corporations are devoting >> > significant time and effort in the project - and these come from several >> > organizations. >> > >> > === A Excessive Fascination with the Apache Brand === >> > >> > We choose Apache not for publicity. We have two purposes. First, we hope >> > that Apache's known best-practices for managing a mature open source >> > project can help guide us. For example, we are feeling the growing pains >> > of a successful open source project as we attempt a major refactor of the >> > internals while customers are using the system in production. We seek >> > guidance in communicating breaking API changes and version revisions. >> > Also, as our involvement from major corporations increases, we want to >> > assure our users that MXNet will stay open and not favor any particular >> > platform or environment. These are some examples of the know-how and >> > discipline we're hoping Apache can bring to our project. >> > >> > Second, we want to leverage Apache's reputation to recruit more >> developers >> > to create a diverse community. >> > >> > === Relationship with Other Apache Products === >> > >> > Apache Mahout and Apache Spark's MLlib are general machine learning >> > systems. Deep learning algorithms can thus be implemented on these two >> > platforms as well. However, in practice, the overlap will be minimal. >> Deep >> > learning is so computationally intensive that it often requires >> specialized >> > GPU hardware to accomplish tasks of meaningful size. Making efficient >> use >> > of GPU hardware is complex because the hardware is so fast that the >> > supporting systems around it must be carefully optimized to keep the GPU >> > cores busy. Extending this capability to distributed multi-GPU and >> > multi-host environments requires great care. This is a critical >> > differentiator between MXNet and existing Apache machine learning >> systems. >> > >> > Mahout and Spark ML-LIB follow models where their nodes run >> synchronously. >> > This is the fundamental difference to MXNet who follows the parameter >> > server framework. MXNet can run synchronously or asynchronously. In >> > addition, MXNet has optimizations for training a wide range of deep >> > learning models using a variety of approaches (e.g., model parallelism >> and >> > data parallelism) which makes MXNet much more efficient (near-linear >> > speedup on state of the art models). MXNet also supports both imperative >> > and symbolic approaches providing ease of programming for deep learning >> > algorithms. >> > >> > Other Apache projects that are potentially complimentary: >> > >> > Apache Arrow - read data in Apache Arrow‘s internal format from MXNet, >> > that would allow users to run ETL/preprocessing in Spark, save the >> results >> > in Arrow’s format and then run DL algorithms on it. >> > >> > Apache Singa - MXNet and Singa are both deep learning projects, and can >> > benefit from a larger deep learning community at Apache. >> > >> > == Documentation == >> > >> > Documentation has recently migrated to http://mxnet.io. We continue to >> > refine and improve the documentation. >> > >> > == Initial Source == >> > >> > We currently use Github to maintain our source code, >> > https://github.com/MXNet >> > >> > == Source and Intellectual Property Submission Plan == >> > >> > MXNet Code is available under Apache License, Version 2.0. We will work >> > with the committers to get CLAs signed and review previous contributions. >> > >> > == External Dependencies == >> > >> > * required by the core code base: GCC or CLOM, Clang, any BLAS library >> > (ATLAS, OpenBLAS, MKL), dmlc-core, mshadow, ps-lite (which requires >> > lib-zeromq), TBB >> > * required for GPU usage: cudnn, cuda >> > * required for python usage: Python 2/3 >> > * required for R module: R, Rcpp (GPLv2 licensing) >> > * optional for image preparation and preprocessing: opencv >> > * optional dependencies for additional features: torch7, numba, cython >> > (in NNVM branch) >> > >> > Rcpt and lib-zeromq are expected to be licensing discussions. >> > >> > == Cryptography == >> > >> > Not Applicable >> > >> > == Required Resources == >> > >> > === Mailing Lists === >> > >> > There is currently no mailing list. >> > >> > === Issue Tracking === >> > >> > Currently uses GitHub to track issues. Would like to continue to do so. >> > >> > == Committers and Affiliations == >> > >> > * Tianqi Chen (UW) >> > * Mu Li (AWS) >> > * Junyuan Xie (AWS) >> > * Bing Xu (Apple) >> > * Chiyuan Zhang (MIT) >> > * Minjie Wang (UYU) >> > * Naiyan Wang (Tusimple) >> > * Yizhi Liu (Mediav) >> > * Tong He (Simon Fraser University) >> > * Qiang Kou (Indiana U) >> > * Xingjian Shi (HKUST) >> > >> > == Sponsors == >> > >> > === Champion === >> > >> > Henri Yandell (bayard at apache.org) >> > >> > === Nominated Mentors === >> > >> > Sebastian Schelter (s...@apache.org) >> > >> > >> > === Sponsoring Entity === >> > >> > We are requesting the Incubator to sponsor this project. >> > >> > >>
-- Yizhi Liu DMLC member Technical Manager Qihoo 360 Inc, Shanghai, China --------------------------------------------------------------------- To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org For additional commands, e-mail: general-h...@incubator.apache.org