Assessing Cloud Network Performance

Performance and scalability of cloud applications are heavily dependent on those same characteristics in the underlying network. Several factors complicate cloud network evaluation. Most notable are 1) large scale, 2) a practically infinite set of application workloads, 3) a large set of potential metrics, 4) a large variety of available configurations within and across clouds and 5) a lack of visibility into the underlying network. There is a tremendous need for a cloud network evaluation methodology that overcomes these challenges, that is practical to apply and provides predictive results. We describe our work to develop such a methodology for cloud network evaluation and a tool (Euclid) that automates its application. Euclid leverages a set of widely used traffic generators to evaluate throughput and latency across a cloud network between endpoint pairs. Evaluation of these micro benchmarks between and among different sets of endpoints is coordinated to reflect several well-known application traffic patterns. Large sets of results are distilled into a tractable set of metrics and associated graphs that facilitate consumption of the results by cloud operators and users. We report results from applying Euclid to seventeen cloud configurations that span three public and one private cloud. The results demonstrate the practicality of our approach and highlight some interesting differences across clouds.

[1]  John Shalf,et al.  Performance Analysis of High Performance Computing Applications on the Amazon Web Services Cloud , 2010, 2010 IEEE Second International Conference on Cloud Computing Technology and Science.

[2]  T. S. Eugene Ng,et al.  Application-specific configuration selection in the cloud: Impact of provider policy and potential of systematic testing , 2015, 2015 IEEE Conference on Computer Communications (INFOCOM).

[3]  Benjamin Farley,et al.  More for your money: exploiting performance heterogeneity in public clouds , 2012, SoCC '12.

[4]  John M. Tracey,et al.  Experiences evaluating OpenStack network data plane performance and scalability , 2016, NOMS 2016 - 2016 IEEE/IFIP Network Operations and Management Symposium.

[5]  Seunghak Lee,et al.  STRADS: a distributed framework for scheduled model parallel machine learning , 2016, EuroSys.

[6]  Dilma Da Silva,et al.  CloudBench: Experiment Automation for Cloud Environments , 2013, 2013 IEEE International Conference on Cloud Engineering (IC2E).

[7]  T. S. Eugene Ng,et al.  The Impact of Virtualization on Network Performance of Amazon EC2 Data Center , 2010, 2010 Proceedings IEEE INFOCOM.

[8]  Antonio Pescapè,et al.  Measuring network throughput in the cloud: The case of Amazon EC2 , 2015, Comput. Networks.

[9]  Daniele Venzano,et al.  A Measurement Study of Data-Intensive Network Traffic Patterns in a Private Cloud , 2013, 2013 IEEE/ACM 6th International Conference on Utility and Cloud Computing.

[10]  Shujia Zhou,et al.  Case study for running HPC applications in public clouds , 2010, HPDC '10.

[11]  Alessio Botta,et al.  On Network Throughput Variability in Microsoft Azure Cloud , 2014, GLOBECOM 2014.

[12]  Xiaowei Yang,et al.  CloudCmp: comparing public cloud providers , 2010, IMC '10.

[13]  Rupak Biswas,et al.  Performance evaluation of Amazon EC2 for NASA HPC applications , 2012, ScienceCloud '12.