Modeling and extracting load intensity profiles

Today's system developers and operators face the challenge of creating software systems that make efficient use of dynamically allocated resources under highly variable and dynamic load profiles, while at the same time delivering reliable performance. Benchmarking of systems under these constraints is difficult, as state-of-the-art benchmarking frameworks provide only limited support for emulating such dynamic and highly variable load profiles for the creation of realistic workload scenarios. Industrial benchmarks typically confine themselves to workloads with constant or stepwise increasing loads. Alternatively, they support replaying of recorded load traces. Statistical load intensity descriptions also do not sufficiently capture concrete pattern load profile variations over time. To address these issues, we present the Descartes Load Intensity Model (DLIM). DLIM provides a modeling formalism for describing load intensity variations over time. A DLIM instance can be used as a compact representation of a recorded load intensity trace, providing a powerful tool for benchmarking and performance analysis. As manually obtaining DLIM instances can be time consuming, we present three different automated extraction methods, which also help to enable autonomous system analysis for self-adaptive systems. Model expressiveness is validated using the presented extraction methods. Extracted DLIM instances exhibit a median modeling error of 12.4% on average over nine different real-world traces covering between two weeks and seven months. Additionally, extraction methods perform orders of magnitude faster than existing time series decomposition approaches.

[1]  Paul Barford,et al.  Generating representative Web workloads for network and server performance evaluation , 1998, SIGMETRICS '98/PERFORMANCE '98.

[2]  Hui Li Realistic Workload Modeling and Its Performance Impacts in Large-Scale eScience Grids , 2010, IEEE Transactions on Parallel and Distributed Systems.

[3]  Dror G. Feitelson,et al.  Workload Modeling for Performance Evaluation , 2002, Performance.

[4]  Adam Wierman,et al.  Open Versus Closed: A Cautionary Tale , 2006, NSDI.

[5]  Dror G. Feitelson,et al.  Workload resampling for performance evaluation of parallel job schedulers , 2014, Concurr. Comput. Pract. Exp..

[6]  Samuel Kounev,et al.  Modeling variations in load intensity over time , 2014, LT '14.

[7]  Jerome A. Rolia,et al.  BURN: Enabling Workload Burstiness in Customized Service Benchmarks , 2012, IEEE Transactions on Software Engineering.

[8]  Rob J Hyndman,et al.  Detecting trend and seasonal changes in satellite image time series , 2010 .

[9]  Samuel Kounev,et al.  LIMBO: a tool for modeling variable load intensities , 2014, ICPE.

[10]  H. Blinchikoff,et al.  Filtering in the time and frequency domains , 1976 .

[11]  Virgílio A. F. Almeida,et al.  A hierarchical and multiscale approach to analyze E-business workloads , 2003, Perform. Evaluation.

[12]  Samuel Kounev,et al.  BUNGEE: An Elasticity Benchmark for Self-Adaptive IaaS Cloud Environments , 2015, 2015 IEEE/ACM 10th International Symposium on Software Engineering for Adaptive and Self-Managing Systems.

[13]  Dominik Benz,et al.  The social bookmark and publication management system bibsonomy , 2010, The VLDB Journal.

[14]  Wilhelm Hasselbring,et al.  Generating Probabilistic and Intensity-Varying Workload for Web-Based Software Systems , 2008, SIPEW.

[15]  Thomas Begin,et al.  A complete framework for modelling and generating workload volatility of a VoD system , 2013, 2013 9th International Wireless Communications and Mobile Computing Conference (IWCMC).

[16]  David A. Patterson,et al.  Rain: A Workload Generation Toolkit for Cloud Computing Applications , 2010 .

[17]  J. Friedman Multivariate adaptive regression splines , 1990 .