Accuracy vs. traffic trade-off of learning IoT data patterns at the edge with hypothesis transfer learning

Right now, the dominant paradigm to support knowledge extraction from raw IoT data is through global cloud platform, where data is collected from IoT devices, and analysed. However, with the ramping trend of the number of IoT devices spread in the physical environment, this approach might simply not scale. The data gravity concept, one of the basis of Fog and Mobile Edge Computing, points towards a decentralisation of computation for data analysis, whereby the latter is performed closer to where data is generated. Along this trend, in this paper we explore the accuracy vs. network traffic trade-off when using Hypothesis Transfer Learning (HTL) to learn patterns from data generated in a set of distributed physical locations. HTL is a standard machine learning technique used to train models on separate disjoint training sets, and then transfer the partial models (instead of the data) to reach a unique learning model. We have previously applied HTL to the problem of learning human activities when data are available in different physical locations (e.g., areas of a city). In our approach, data is not moved from where it is generated, while partial models are exchanged across sites. The HTL-based approach achieves lower (though acceptable) accuracy with respect to a conventional solution based on global cloud computing, but drastically cuts the network traffic. In this paper we explore the trade-off between accuracy and traffic, by assuming that data are moved to a variable number of data collectors where partial learning is performed. Centralised cloud and completely decentralised HTL are the two extremes of the spectrum. Our results show that there is no significant advantage in terms of accuracy, in using fewer collectors, and that therefore a distributed HTL solutions, along the lines of a fog computing approach, is the most promising one.

[1]  Tao Wang,et al.  Deep learning with COTS HPC systems , 2013, ICML.

[2]  Dianhui Wang,et al.  Distributed learning for Random Vector Functional-Link networks , 2015, Inf. Sci..

[3]  Martin Hasler,et al.  Distributed machine learning in networks by consensus , 2014, Neurocomputing.

[4]  Jason Weston,et al.  Fast Kernel Classifiers with Online and Active Learning , 2005, J. Mach. Learn. Res..

[5]  Marco Conti,et al.  Hypothesis Transfer Learning for Efficient Data Computing in Smart Cities Environments , 2016, 2016 IEEE International Conference on Smart Computing (SMARTCOMP).

[6]  Ilja Kuzborskij,et al.  Transfer Learning Through Greedy Subset Selection , 2014, ICIAP.

[7]  Teruo Higashino,et al.  Edge-centric Computing: Vision and Challenges , 2015, CCRV.

[8]  Marc'Aurelio Ranzato,et al.  Large Scale Distributed Deep Networks , 2012, NIPS.

[9]  Yoram Singer,et al.  Using and combining predictors that specialize , 1997, STOC '97.

[10]  Gary M. Weiss,et al.  Design considerations for the WISDM smart phone-based sensor mining architecture , 2011, SensorKDD '11.

[11]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[12]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[13]  Geoffrey E. Hinton,et al.  Adaptive Mixtures of Local Experts , 1991, Neural Computation.

[14]  J. Manyika,et al.  Disruptive technologies: Advances that will transform life, business, and the global economy , 2013 .

[15]  Emilio Parrado-Hernández,et al.  Distributed support vector machines , 2006, IEEE Trans. Neural Networks.

[16]  Eleonora Borgia,et al.  The Internet of Things vision: Key features, applications and open issues , 2014, Comput. Commun..

[17]  Davide Anguita,et al.  Transition-Aware Human Activity Recognition Using Smartphones , 2016, Neurocomputing.