Inferring Origin Wi-Fi Traffic Patterns by Deep Learning of Max-pooled Sparse Representations of Received Frames

We introduce a novel networking application of deep learning to infer the origin traffic pattern in Wi-Fi from observing only received frame runs and gaps. CSMA adds linear and nonlinear altercations to the original characteristics of a flow, which is subject to further distortions from mixing with other flows. We devise a multi-layer inference pipeline based on sparse coding and dictionary learning popular in machine learning. In particular, we integrate max pooling at all layers to capture features (received frame characteristics) varying in length and position. We empirically evaluate the proposed pipeline in OPNET and present a comparative performance analysis with ARX/ARMAX least squares, Naı̈ve Bayes classifier, kernel density estimation, and EM-optimized Gaussian mixture model.

[1]  Kevin Chen,et al.  Wireless Inference-based Notification (WIN) without Packet Decoding , 2013, ICAC.

[2]  Hongbin Zha,et al.  Incoherent dictionary learning for sparse representation , 2012, Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012).

[3]  Andrew Y. Ng,et al.  Selecting Receptive Fields in Deep Networks , 2011, NIPS.

[4]  Dario Vlah,et al.  A location-dependent runs-and-gaps model for predicting TCP performance over a UAV wireless channel , 2010, 2010 - MILCOM 2010 MILITARY COMMUNICATIONS CONFERENCE.

[5]  Baoxin Li,et al.  Discriminative K-SVD for dictionary learning in face recognition , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[6]  Guillermo Sapiro,et al.  Classification and clustering via dictionary learning with structured incoherence and shared features , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[7]  Allen Y. Yang,et al.  Robust Face Recognition via Sparse Representation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[8]  Rajat Raina,et al.  Self-taught learning: transfer learning from unlabeled data , 2007, ICML '07.

[9]  Jan A Snyman,et al.  Practical Mathematical Optimization: An Introduction to Basic Optimization Theory and Classical and New Gradient-Based Algorithms , 2005 .

[10]  A. Bruckstein,et al.  K-SVD : An Algorithm for Designing of Overcomplete Dictionaries for Sparse Representation , 2005 .

[11]  R. Tibshirani,et al.  Least angle regression , 2004, math/0406456.

[12]  Tomaso A. Poggio,et al.  Face recognition with support vector machines: global versus component-based approach , 2001, Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001.

[13]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[14]  T. Poggio,et al.  Hierarchical models of object recognition in cortex , 1999, Nature Neuroscience.

[15]  Lennart Ljung,et al.  System identification (2nd ed.): theory for the user , 1999 .

[16]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[17]  Pat Langley,et al.  Estimating Continuous Distributions in Bayesian Classifiers , 1995, UAI.

[18]  Y. C. Pati,et al.  Orthogonal matching pursuit: recursive function approximation with applications to wavelet decomposition , 1993, Proceedings of 27th Asilomar Conference on Signals, Systems and Computers.

[19]  Bernhard E. Boser,et al.  A training algorithm for optimal margin classifiers , 1992, COLT '92.

[20]  T. Kailath,et al.  Fast, recursive-least-squares transversal filters for adaptive filtering , 1984 .

[21]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .