Ease.ml/ci and Ease.ml/meter in Action: Towards Data Management for Statistical Generalization

Developing machine learning (ML) applications is similar to developing traditional software — it is often an iterative process in which developers navigate within a rich space of requirements, design decisions, implementations, empirical quality, and performance. In traditional software development, software engineering is the field of study which provides principled guidelines for this iterative process. However, as of today, the counterpart of “software engineering for ML” is largely missing — developers of ML applications are left with powerful tools (e.g., TensorFlow and PyTorch) but little guidance regarding the development lifecycle itself. In this paper, we view the management of ML development lifecycles from a data management perspective. We demonstrate two closely related systems, ease.ml/ci and ease.ml/meter,  that provide some “principled guidelines” for ML application development: ci is a continuous integration engine for ML models and meter is a “profiler” for controlling overfitting of ML models. Both systems focus on managing the “statistical generalization power” of datasets used for assessing the quality of ML applications, namely, the validation set and the test set. By demonstrating these two systems we hope to spawn further discussions within our community on building this new type of data management systems for statistical generalization.

[1]  Ce Zhang,et al.  Using transfer learning to detect galaxy mergers , 2018, Monthly Notices of the Royal Astronomical Society.

[2]  Toniann Pitassi,et al.  Preserving Statistical Validity in Adaptive Data Analysis , 2014, STOC.

[3]  Avrim Blum,et al.  The Ladder: A Reliable Leaderboard for Machine Learning Competitions , 2015, ICML.

[4]  Nora Hollenstein,et al.  Patient Risk Assessment and Warning Symptom Detection Using Deep Attention-Based Neural Networks , 2018, Louhi@EMNLP.

[5]  Tian Li,et al.  Ease.ml: Towards Multi-tenant Resource Sharing for Machine Learning Workloads , 2017, Proc. VLDB Endow..

[6]  Ce Zhang,et al.  Exploring galaxy evolution with generative models , 2018, Astronomy & Astrophysics.

[7]  Dan Alistarh,et al.  ZipML: Training Linear Models with End-to-End Low Precision, and a Little Bit of Deep Learning , 2017, ICML.

[8]  Feng Liu,et al.  Continuous Integration of Machine Learning Models with ease.ml/ci: Towards a Rigorous Yet Practical Treatment , 2019, SysML.

[9]  Nora Hollenstein,et al.  ETH-DS3Lab at SemEval-2018 Task 7: Effectively Combining Recurrent and Convolutional Neural Networks for Relation Classification and Extraction , 2018, *SEMEVAL.

[10]  Alexander J. Smola,et al.  Scaling Distributed Machine Learning with the Parameter Server , 2014, OSDI.

[11]  Kevin Schawinski,et al.  psfgan: a generative adversarial network system for separating quasar point sources and host galaxy light , 2018, Monthly Notices of the Royal Astronomical Society.

[12]  Ce Zhang,et al.  Ease.ml/meter: Quantitative Overfitting Management for Human-in-the-loop ML Application Development , 2019 .

[13]  Min Su,et al.  Generative adversarial networks as a tool to recover structural information from cryo-electron microscopy data , 2018, bioRxiv.

[14]  Ce Zhang,et al.  Generative Adversarial Networks recover features in astrophysical images of galaxies beyond the deconvolution limit , 2017, ArXiv.

[15]  Tim Kraska,et al.  Northstar: An Interactive Data Science System , 2018, Proc. VLDB Endow..