E2Clab: Exploring the Computing Continuum through Repeatable, Replicable and Reproducible Edge-to-Cloud Experiments

Distributed digital infrastructures for computation and analytics are now evolving towards an interconnected ecosystem allowing complex applications to be executed from IoT Edge devices to the HPC Cloud (aka the Computing Continuum, the Digital Continuum, or the Transcontinuum). Understanding end-to-end performance in such a complex continuum is challenging. This breaks down to reconciling many, typically contradicting application requirements and constraints with low-level infrastructure design choices. One important challenge is to accurately reproduce relevant behaviors of a given application workflow and representative settings of the physical infrastructure underlying this complex continuum. In this paper we introduce a rigorous methodology for such a process and validate it through E2Clab. It is the first platform to support the complete analysis cycle of an application on the Computing Continuum: (i) the configuration of the experimental environment, libraries and frameworks; (ii) the mapping between the application parts and machines on the Edge, Fog and Cloud; (iii) the deployment of the application on the infrastructure; (iv) the automated execution; and (v) the gathering of experiment metrics. We illustrate its usage with a real-life application deployed on the Grid’5000 testbed, showing that our framework allows one to understand and improve performance, by correlating it to the parameter settings, the resource usage and the specifics of the underlying infrastructure.

[1]  Gabriel Antoniu,et al.  Investigating Edge vs. Cloud Computing Trade-offs for Stream Processing , 2019, 2019 IEEE International Conference on Big Data (Big Data).

[2]  Dimitrios Tzovaras,et al.  Simulating Fog and Edge Computing Scenarios: An Overview and Research Challenges , 2019, Future Internet.

[3]  Victoria Stodden,et al.  Best Practices for Computational Science: Software Infrastructure and Environments for Reproducible and Extensible Research , 2014 .

[4]  Blesson Varghese,et al.  DeFog: fog computing benchmarks , 2019, SEC.

[5]  Scott Shenker,et al.  Spark: Cluster Computing with Working Sets , 2010, HotCloud.

[6]  Maria E. Orlowska,et al.  Data Flow and Validation in Workflow Modelling , 2004, ADC.

[7]  Nicola Ferro,et al.  SIGIR Initiative to Implement ACM Artifact Review and Badging , 2018, SIGF.

[8]  Xiaowei Yang,et al.  CloudCmp: comparing public cloud providers , 2010, IMC '10.

[9]  Matthieu Simonin,et al.  EnosStack: A LAMP-like stack for the experimenter , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[10]  George K. Thiruvathukal,et al.  Reproducible Research for Computing in Science & Engineering , 2017, Comput. Sci. Eng..

[11]  Stacy Patterson,et al.  EdgeBench: Benchmarking Edge Computing Platforms , 2018, 2018 IEEE/ACM International Conference on Utility and Cloud Computing Companion (UCC Companion).

[12]  Franck Cappello,et al.  Big data and extreme-scale computing , 2018, Int. J. High Perform. Comput. Appl..

[13]  Henri E. Bal,et al.  SWAN-Fly : A flexible cloud-enabled framework for context-aware applications in smartphones , 2016 .

[14]  Franck Cappello,et al.  Grid'5000: A Large Scale And Highly Reconfigurable Experimental Grid Testbed , 2006, Int. J. High Perform. Comput. Appl..

[15]  Nuno Vasconcelos,et al.  Modeling, Clustering, and Segmenting Video with Mixtures of Dynamic Textures , 2008, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[16]  Karima Velasquez,et al.  A comparative analysis of simulators for the Cloud to Fog continuum , 2020, Simul. Model. Pract. Theory.

[17]  Paul Rad,et al.  Chameleon: A Scalable Production Testbed for Computer Science Research , 2019, Contemporary High Performance Computing.

[18]  Seif Haridi,et al.  Apache Flink™: Stream and Batch Processing in a Single Engine , 2015, IEEE Data Eng. Bull..

[19]  Robert Barton,et al.  Fog Computing Conceptual Model , 2018 .

[20]  Wolfgang Kellerer,et al.  NCSbench: Reproducible Benchmarking Platform for Networked Control Systems , 2020, 2020 IEEE 17th Annual Consumer Communications & Networking Conference (CCNC).