Sub projects are frowned upon. It is possible for a project to graduate as part of another project from. The incubator, but that is very unusual. Graduating to a very quiet project like hama would be even more unusual.
A better course would be to simply create a new incubator project. Worry about building a viable project first. Worry about graduation details later. On Mar 2, 2017 7:35 AM, "Edward Capriolo" <edlinuxg...@gmail.com> wrote: > On Mon, Feb 27, 2017 at 7:13 PM, Edward J. Yoon <edward.y...@samsung.com> > wrote: > > > Thanks for your proposal. > > > > I of course think Apache Hama can be used for scheduling sync and async > > communication/computation networks with various topologies and resource > > allocation. However, I'm not sure whether this approach is also fit for > > modern microservice architecture? In my opinion, this can be discussed > and > > cooked in Hama community as a sub-project until it's mature enough > (CC'ing > > general@i.a.o. I'll be happy to read more feedbacks from ASF incubator > > community). > > > > P.S., It seems you referred to incubation proposal template. There's no > > need > > to add me as initial committer (I don't have much time to actively > > contribute to your project). And, I recently quit Samsung Electronics and > > joined to $200 billion sized O2O e-commerce company as a CTO. > > > > -----Original Message----- > > From: Sachin Ghai [mailto:sachin.g...@impetus.co.in] > > Sent: Monday, February 27, 2017 5:16 PM > > To: d...@hama.apache.org > > Subject: Proposal for an Apache Hama sub-project > > > > Hama Community, > > > > I would like to propose a sub-project for Apache Hama and initiate > > discussion around the proposal. The proposed sub-project named 'Scalar' > is > > a > > scalable orchestration, training and serving system for machine learning > > and > > deep learning. Scalar would leverage Apache Hama to automate the > > distributed > > training, model deployment and prediction serving. > > > > More details about the proposal are listed below as per Apache project > > proposal template: > > Abstract > > Scalar is a general purpose framework for simplifying massive scale big > > data > > analytics and deep learning modelling, deployment, serving with high > > performance. > > Proposal > > It is a goal of Scalar to provide an abstraction framework which allows > > user > > to easily scale the functions of training a model, deploying a model and > > serving the prediction from underlying machine learning or deep learning > > framework. It is also the characteristic of its execution framework to > > orchestrate heterogeneous workload graphs utilizing Apache Hama, Apache > > Hadoop, Apache Spark and TensorFlow resources. > > Background > > The initial Scalar code was developed in 2016 and has been successfully > > beta > > tested for one of the largest insurance organizations in a client > specific > > PoC. The motivation behind this work is to build a framework that > provides > > abstraction on heterogeneous data science frameworks and helps users > > leverage them in the most performant way. > > Rationale > > There is a sudden deluge of machine learning and deep learning frameworks > > in > > the industry. As an application developer, it becomes a hard choice to > > switch from one framework to another without rewriting the application. > > Also, there is additional plumbing to be done to retrieve the prediction > > results for each model in different frameworks. We aim to provide an > > abstraction framework which can be used to seamlessly train and deploy > the > > model at scale on multiple frameworks like TensorFlow, Apache Horn or > > Caffe. > > The abstraction further provides a unified layer for serving the > prediction > > in the most performant, scalable and efficient way for a multi-tenant > > deployment. The key performance metrics will be reduction in training > time, > > lower error rate and lower latency time for serving models. > > Scalar consists of a core engine which can be used to create flows > > described > > in terms of state, sequences and algorithms. The engine invokes execution > > context of Apache Hama to train and deploy models on target framework. > > Apache Hama is used for a variety of functions including parameter tuning > > and scheduling computations on a distributed cluster. A data object layer > > provides access to data from heterogeneous sources like HDFS, local, S3 > > etc. > > A REST API layer is utilized for serving the prediction functions to > client > > applications. A caching layer in the middle acts as a latency improver > for > > various functions. > > Initial Goals > > Some current goals include: > > > > * Build community. > > * Provide general purpose API for machine learning and deep learning > > training, deployment and serving. > > * Serve the predictions with low latency. > > * Run massive workloads via Apache Hama on TensorFlow, Apache Spark > and > > Caffe. > > * Provide CPU and GPU support on-premise or on cloud to run the > > algorithms. > > Current Status > > Meritocracy > > The core developers understand what it means to have a process based on > > meritocracy. We will provide continuous efforts to build an environment > > that > > supports this, encouraging community members to contribute. > > Community > > A small community has formed within the Apache Hama project community and > > companies such as enterprise services and product company and artificial > > intelligence startup. There is a lot of interest in data science serving > > systems and Artificial intelligence simplification systems. By bringing > > Scalar into Apache, we believe that the community will grow even bigger. > > Core Developers > > Edward J. Yoon, Sachin Ghai, Ishwardeep Singh, Rachna Gogia, Abhishek > Soni, > > Nikunj Limbaseeya, Mayur Choubey > > Known Risks > > Orphaned Products > > Apache Hama is already a core open source component being utilized at > > Samsung Electronics, and Scalar is already getting adopted by major > > enterprise organizations. There is no direct risk for Scalar project to > be > > orphaned. > > Inexperience with Open Source > > All contributors have experience using and/or working on Apache open > source > > projects. > > Homogeneous Developers > > The initial committers are from different organizations such as Impetus, > > Chalk Digital, and Samsung Electronics. > > Reliance on Salaried Developers > > Few will be working as full-time open source developer. Other developers > > will also start working on the project in their spare time. > > Relationships with Other Apache Products > > > > * Scalar is being built on top of Apache Hama > > * Apache Spark is being used for machine learning. > > * Apache Horn is being used for deep learning. > > * The framework will run natively on Apache Hadoop and Apache Mesos. > > An Excessive Fascination with the Apache Brand > > Scalar itself will hopefully have benefits from Apache, in terms of > > attracting a community and establishing a solid group of developers, but > > also the relation with Apache Hadoop, Spark and Hama. These are the main > > reasons for us to send this proposal. > > Documentation > > Initial design of Scalar can be found at this > > link<https://drive.google.com/file/d/0B7mbLUemi6LFVHlFSzhONm > > Z4aU0/view?usp=s > > haring>. > > Initial Source > > Impetus Technologies (Impetus) will contribute the initial orchestration > > code base to create this project. Impetus plans to contribute the Scalar > > code base, test cases, build files, and documentation to the ASF under > the > > terms specified in the ASF Corporate Contributor License and further > > develop > > it with wider community. Once at Apache, the project will be licensed > under > > the ASF license. > > Cryptography > > Not applicable. > > Required Resources > > Mailing Lists > > > > * scalar-dev > > * scalar-pmc > > Subversion Directory > > > > * Git is the preferred source control system: > > git://git.apache.org/scalar > > Issue Tracking > > > > * a JIRA issue tracker, SCALAR > > Initial Committers > > > > * Sachin Ghai (sachin.ghai AT impetus DOT co DOT in) > > * Edward J. Yoon (edwardyoon AT apache DOT org) > > * Abhishek Soni (abhishek.soni AT impetus DOT co DOT in) > > * Ishwardeep Singh ( ishwardeep AT chalkdigital DOT com ) > > * Nikunj Limbaseeya (nikunj.limbaseeya AT impetus DOT co DOT in) > > * Rachna Gogia (rachna AT hadoopsphere DOT org) > > * Mayur Choubey (mayur.choubey AT impetus DOT co DOT in) > > Affiliations > > > > * Sachin Ghai (Impetus) > > * Edward J. Yoon (Samsung Electronics) > > * Abhishek Soni (Impetus) > > * Ishwardeep Singh ( Chalk Digital) > > * Nikunj Limbaseeya (Impetus) > > * Rachna Gogia (HadoopSphere) > > * Mayur Choubey (Impetus) > > Sponsors > > <proposed> > > Champion > > > > * Edward J. Yoon <ASF member, Samsung Electronics > > > Nominated Mentors > > > > * Edward J. Yoon <ASF member, Samsung Electronics > > > Sponsoring Entity > > The Apache Hama project > > > > -- End of proposal -- > > > > Thanks, > > Sachin Ghai > > > > ________________________________ > > > > > > > > > > > > > > NOTE: This message may contain information that is confidential, > > proprietary, privileged or otherwise protected by law. The message is > > intended solely for the named addressee. If received in error, please > > destroy and notify the sender. Any use of this email is prohibited when > > received in error. Impetus does not represent, warrant and/or guarantee, > > that the integrity of this communication has been maintained nor that the > > communication is free of errors, virus, interception or interference. > > > > > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org > > For additional commands, e-mail: general-h...@incubator.apache.org > > > > > I do not believe the the Hama project has had activity for a long time. 1 + > year. For example, have attempted to broach this discussion and got no > official reply: https://issues.apache.org/jira/browse/HAMA-998. > > I am interested in Scalar and I would like to take time and familiarize > myself with it. I do not believe I am the right champion but I can > possibly be a mentor/contributor. > > Thanks, > Edward >