DEEP-Mon: Dynamic and Energy Efficient Power Monitoring for Container-Based Infrastructures

In the last few years energy efficiency of large scale infrastructures gained a lot of attention, as power consumption became one of the most impacting factors of the operative costs of a data-center and of its Total Cost of Ownership (TCO). Power consumption can be observed at different layers of the data-center, from the overall power grid, moving to each rack and arriving to each machine and system. Given the rise of application containers both in the cloud computing and High Performance Computing (HPC) scenarios, it becomes more and more important to measure power consumption also at the application level, where power-aware schedulers and orchestrators can optimize the execution of the workloads not only from a performance perspective, but also considering performance/power trade-offs. In this paper we propose DEEP-mon, a novel monitoring tool able to measure power consumption and attribute it for each thread and application container running in the system. Moreover, we show how the proposed approach has a negligible impact on the monitored system and on the running workloads, overcoming the limitations of the previous works in the field.

[1]  John Kubiatowicz,et al.  Power Consumption Models for Multi-Tenant Server Infrastructures , 2017, ACM Trans. Archit. Code Optim..

[2]  Wolf-Dietrich Weber,et al.  Power provisioning for a warehouse-sized computer , 2007, ISCA '07.

[3]  David H. Bailey,et al.  The Nas Parallel Benchmarks , 1991, Int. J. High Perform. Comput. Appl..

[4]  Marco D. Santambrogio,et al.  DockerCap: A Software-Level Power Capping Orchestrator for Docker Containers , 2016, 2016 IEEE Intl Conference on Computational Science and Engineering (CSE) and IEEE Intl Conference on Embedded and Ubiquitous Computing (EUC) and 15th Intl Symposium on Distributed Computing and Applications for Business Engineering (DCABES).

[5]  Maximilien de Bayser,et al.  Integrating MPI with Docker for HPC , 2017, 2017 IEEE International Conference on Cloud Engineering (IC2E).

[6]  Steven McCanne,et al.  The BSD Packet Filter: A New Architecture for User-level Packet Capture , 1993, USENIX Winter.

[7]  John Kubiatowicz,et al.  Enabling power-awareness for the Xen hypervisor , 2018, SIGBED.

[8]  Christoforos E. Kozyrakis,et al.  Towards energy proportionality for large-scale latency-critical workloads , 2014, 2014 ACM/IEEE 41st International Symposium on Computer Architecture (ISCA).

[9]  Kevin T. Pedretti,et al.  A Tale of Two Systems: Using Containers to Deploy HPC Applications on Supercomputers and Clouds , 2017, 2017 IEEE International Conference on Cloud Computing Technology and Science (CloudCom).

[10]  Rajkumar Buyya,et al.  A Framework and Algorithm for Energy Efficient Container Consolidation in Cloud Data Centers , 2015, 2015 IEEE International Conference on Data Science and Data Intensive Systems.

[11]  Frank Bellosa,et al.  The benefits of event: driven energy accounting in power-sensitive systems , 2000, ACM SIGOPS European Workshop.

[12]  Sriram Sankar,et al.  The need for speed and stability in data center power capping , 2012, 2012 International Green Computing Conference (IGCC).

[13]  Eric A. Brewer,et al.  Borg, Omega, and Kubernetes , 2016, ACM Queue.

[14]  Emily Le,et al.  Performance Analysis of Applications using Singularity Container on SDSC Comet , 2017, PEARC.

[15]  Xiao Zhang,et al.  Power containers: an OS facility for fine-grained power and energy management on multicore servers , 2013, ASPLOS '13.

[16]  John Kubiatowicz,et al.  MARC: A Resource Consumption Modeling Service for Self-Aware Autonomous Agents , 2018, ACM Trans. Auton. Adapt. Syst..

[17]  Ghaleb Abdulla,et al.  Towards a Unified Monitoring Framework for Power, Performance and Thermal Metrics: A Case Study on the Evaluation of HPC Cooling Systems , 2017, 2017 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW).

[18]  Rajkumar Buyya,et al.  CloudSim: a toolkit for modeling and simulation of cloud computing environments and evaluation of resource provisioning algorithms , 2011, Softw. Pract. Exp..

[19]  村井 均,et al.  NAS Parallel Benchmarks によるHPFの評価 , 2006 .

[20]  Albert Y. Zomaya,et al.  A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems , 2010, Adv. Comput..

[21]  Steven McCanne,et al.  BPF+: exploiting global data-flow optimization in a generalized packet filter architecture , 1999, SIGCOMM '99.

[22]  Efraim Rotem,et al.  Power-Management Architecture of the Intel Microarchitecture Code-Named Sandy Bridge , 2012, IEEE Micro.