Performance Analysis of Open Source Machine Learning Frameworks for Various Parameters in Single-Threaded and Multi-Threaded Modes

The basic features of some of the most versatile and popular open source frameworks for machine learning (TensorFlow, Deep Learning4j, and H2O) are considered and compared. Their comparative analysis was performed and conclusions were made as to the advantages and disadvantages of these platforms. The performance tests for the de facto standard MNIST data set were carried out on H2O framework for deep learning algorithms designed for CPU and GPU platforms for single-threaded and multithreaded modes of operation Also, we present the results of testing neural networks architectures on H2O platform for various activation functions, stopping metrics, and other parameters of machine learning algorithm. It was demonstrated for the use case of MNIST database of handwritten digits in single-threaded mode that blind selection of these parameters can hugely increase (by 2-3 orders) the runtime without the significant increase of precision. This result can have crucial influence for optimization of available and new machine learning methods, especially for image recognition problems.

[1]  Yuri Gordienko,et al.  Comparative performance analysis of neural networks architectures on H2O platform for various activation functions , 2017, 2017 IEEE International Young Scientists Forum on Applied Physics and Engineering (YSF).

[2]  Ugo Becciani,et al.  VO-compliant workflows and science gateways , 2015, Astron. Comput..

[3]  Gábor Terstyánszky,et al.  Quantum chemical meta‐workflows in MoSGrid , 2015, Concurr. Comput. Pract. Exp..

[4]  Sergii Stirenko,et al.  IMP Science Gateway: from the Portal to the Hub of Virtual Experimental Labs in e‐Science and Multiscale Courses in e‐Learning , 2015, Concurr. Comput. Pract. Exp..

[5]  Yuri Gordienko,et al.  Deep Learning for Fatigue Estimation on the Basis of Multimodal Human-Machine Interactions , 2017, ArXiv.

[6]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[7]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[8]  อนิรุธ สืบสิงห์,et al.  Data Mining Practical Machine Learning Tools and Techniques , 2014 .

[9]  Douglas B. Kell,et al.  Software review: the KNIME workflow environment and its applications in genetic programming and machine learning , 2015, Genetic Programming and Evolvable Machines.

[10]  Yuri Gordienko,et al.  User-driven Intelligent Interface on the Basis of Multimodal Augmented Reality and Brain-Computer Interaction for People with Functional Disabilities , 2017, Advances in Intelligent Systems and Computing.

[11]  Yuri Gordienko,et al.  Comparative analysis of open source frameworks for machine learning with use case in single-threaded and multi-threaded modes , 2017, 2017 12th International Scientific and Technical Conference on Computer Sciences and Information Technologies (CSIT).

[12]  K. Skala,et al.  Augmented Coaching Ecosystem for Non-obtrusive Adaptive Personalized Elderly Care on the basis of Cloud-Fog-Dew computing paradigm , 2017, 2017 40th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO).

[13]  Ian H. Witten,et al.  Chapter 10 – Deep learning , 2017 .

[14]  Yuri Gordienko,et al.  Automatized generation of alphabets of symbols , 2017, 2017 Federated Conference on Computer Science and Information Systems (FedCSIS).

[15]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[16]  Judy Qiu,et al.  Software Frameworks for Deep Learning at Scale , 2016 .

[17]  Miklós Kozlovszky,et al.  DCI Bridge: Executing WS-PGRADE Workflows in Distributed Computing Infrastructures , 2014, Science Gateways for Distributed Computing Infrastructures.