Data mining analysis to validate performance tuning practices for HPL

Applications performance is a criterion for system evaluation, and hence performance tuning for these applications is of great interest. One such benchmark application is High Performance Linpack (HPL). Although guidelines exist for HPL tuning, validating these guidelines on various systems is a challenging task as a large number of configurations need to be tested. In this work, we use data mining analysis to reduce the number of configurations to be tested in validating the HPL tuning guidelines on the Ranger System. We validate that NB, P and Q are the three most important parameters to tune HPL, and that PMAP does not have a significant impact on HPL performance. We also validate the practice of tuning HPL at small N using data mining analysis. We find that the value of N selected for tuning should not be significantly smaller than the largest N that can fit into the system memory. Our results indicate that data mining could be further applied to application performance tuning.

[1]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[2]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[3]  Shuen-Tai Wang,et al.  A semi-empirical model for maximal LINPACK performance predictions , 2006, Sixth IEEE International Symposium on Cluster Computing and the Grid (CCGRID'06).

[4]  Yuichi Inadomi,et al.  Performance prediction of large-scale parallell system and application using macro-level simulation , 2008, 2008 SC - International Conference for High Performance Computing, Networking, Storage and Analysis.

[5]  William J. Dally,et al.  Principles and Practices of Interconnection Networks , 2004 .

[6]  Raj Jain,et al.  The Art of Computer Systems Performance Analysis : Tech-niques for Experimental Design , 1991 .

[7]  A. Arnold,et al.  Harvesting graphics power for MD simulations , 2007, 0709.3225.

[9]  Kuo-Chan Huang,et al.  An Improved Model for Predicting HPL Performance , 2007, GPC.

[10]  Yuichi Inadomi,et al.  Performance prediction of large-scale parallell system and application using macro-level simulation , 2008, HiPC 2008.

[11]  Ian Witten,et al.  Data Mining , 2000 .

[12]  Sally A. McKee,et al.  Methods of inference and learning for performance modeling of parallel applications , 2007, PPoPP.

[13]  Minh-Nghia Nguyen,et al.  Design, Deployment and Bench of a Large Infiniband HPC Cluster , 2006, 20th International Symposium on High-Performance Computing in an Advanced Collaborative Environment (HPCS'06).

[14]  Leonid Oliker,et al.  Reconfigurable hybrid interconnection for static and dynamic scientific applications , 2007, CF '07.

[15]  Kenichi Miura,et al.  Performance Improvement Methodology for ClearSpeed's CSX600 , 2007, 2007 International Conference on Parallel Processing (ICPP 2007).

[16]  Raj Jain,et al.  The art of computer systems performance analysis - techniques for experimental design, measurement, simulation, and modeling , 1991, Wiley professional computing.

[17]  Feng-Hsiung Hsu,et al.  Behind Deep Blue: Building the Computer that Defeated the World Chess Champion , 2002 .

[18]  Sally A. McKee,et al.  Predicting parallel application performance via machine learning approaches , 2007, Concurr. Comput. Pract. Exp..

[19]  Joshua A. Anderson,et al.  General purpose molecular dynamics simulations fully implemented on graphics processing units , 2008, J. Comput. Phys..

[20]  David G. Stork,et al.  Pattern Classification , 1973 .