Fast Adaptation of Activity Sensing Policies in Mobile Devices

With the proliferation of sensors, such as accelerometers, in mobile devices, activity and motion tracking has become a viable technology to understand and create an engaging user experience. This paper proposes a fast adaptation and learning scheme of activity tracking policies when user statistics are unknown a priori, varying with time, and inconsistent for different users. In our stochastic optimization, user activities are required to be synchronized with a backend under a cellular data limit to avoid overcharges from cellular operators. The mobile device is charged intermittently using wireless or wired charging for receiving the required energy for transmission and sensing operations. First, we propose an activity tracking policy by formulating a stochastic optimization as a constrained Markov decision process (CMDP). Second, we prove that the optimal policy of the CMDP has a threshold structure using a Lagrangian relaxation approach and the submodularity concept. We accordingly present a fast Q-learning algorithm by considering the policy structure to improve the convergence speed over that of conventional Q-learning. Finally, simulation examples are presented to support the theoretical findings of this paper.

[1]  D. M. Topkis Supermodularity and Complementarity , 1998 .

[2]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.

[3]  Vikram Krishnamurthy,et al.  MIMO Transmission Control in Fading Channels—A Constrained Markov Decision Process Formulation With Monotone Randomized Policies , 2007, IEEE Transactions on Signal Processing.

[4]  Bhaskar Krishnamachari,et al.  Markov-optimal sensing policy for user state estimation in mobile devices , 2010, IPSN '10.

[5]  Chuan Qin,et al.  TagSense: Leveraging Smartphones for Automatic Image Tagging , 2014, IEEE Transactions on Mobile Computing.

[6]  R. Syski,et al.  Fundamentals of Queueing Theory , 1999, Technometrics.

[7]  Arkady B. Zaslavsky,et al.  Context Aware Computing for The Internet of Things: A Survey , 2013, IEEE Communications Surveys & Tutorials.

[8]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[9]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[10]  Linn I. Sennott,et al.  Constrained Average Cost Markov Decision Chains , 1993, Probability in the Engineering and Informational Sciences.

[11]  Hojung Cha,et al.  SmartDC: Mobility Prediction-Based Adaptive Duty Cycling for Everyday Location Monitoring , 2014, IEEE Transactions on Mobile Computing.

[12]  Tao Feng,et al.  TIPS: context-aware implicit user identification using touch screen in uncontrolled environments , 2014, HotMobile.

[13]  Miguel A. Labrador,et al.  A Survey on Human Activity Recognition using Wearable Sensors , 2013, IEEE Communications Surveys & Tutorials.

[14]  F. Beutler,et al.  Optimal policies for controlled markov chains with a constraint , 1985 .

[15]  Vikram Krishnamurthy,et al.  Monotonicity of Constrained Optimal Transmission Policies in Correlated Fading Channels With ARQ , 2010, IEEE Transactions on Signal Processing.

[16]  Johannes Peltola,et al.  Activity classification using realistic data from wearable sensors , 2006, IEEE Transactions on Information Technology in Biomedicine.

[17]  Ozgur Yurur,et al.  Adaptive and Energy Efficient Context Representation Framework in Mobile Sensing , 2014, IEEE Transactions on Mobile Computing.

[18]  E. Altman Constrained Markov Decision Processes , 1999 .

[19]  Emil Jovanov,et al.  APPLICATIONS OF SMARTPHONES FOR UBIQUITOUS HEALTH MONITORING AND WELLBEING MANAGEMENT , 2011 .

[20]  Zhigang Liu,et al.  The Jigsaw continuous sensing engine for mobile phone applications , 2010, SenSys '10.

[21]  Yi Wang,et al.  A framework of energy efficient mobile sensing for automatic user state recognition , 2009, MobiSys '09.

[22]  Dina Katabi,et al.  Magnetic MIMO: how to charge your phone in your pocket , 2014, MobiCom.

[23]  Lixin Shi,et al.  Wireless Power Hotspot that Charges All of Your Devices , 2015, MobiCom.

[24]  Leslie Pack Kaelbling,et al.  On the Complexity of Solving Markov Decision Problems , 1995, UAI.

[25]  Araceli Sanchis,et al.  Activity Recognition Using Hybrid Generative/Discriminative Models on Home Environments Using Binary Sensors , 2013, Sensors.

[26]  Angela Doufexi,et al.  Hotspot wireless LANs to enhance the performance of 3G and beyond cellular networks , 2003, IEEE Communications Magazine.

[27]  Hwee Pink Tan,et al.  Mobile big data analytics using deep learning and apache spark , 2016, IEEE Network.

[28]  Hojung Cha,et al.  Adaptive Duty Cycling for Place-Centric Mobility Monitoring using Zero-Cost Information in Smartphone , 2014, IEEE Transactions on Mobile Computing.

[29]  Min Chen,et al.  Energy-Efficient and Context-Aware Smartphone Sensor Employment , 2015, IEEE Transactions on Vehicular Technology.

[30]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[31]  James Begole,et al.  An Exploration into Activity-Informed Physical Advertising Using PEST , 2007, Pervasive.

[32]  Daniel P. Heyman,et al.  Stochastic models in operations research , 1982 .

[33]  Carl M. Harris,et al.  Fundamentals of Queueing Theory: Gross/Fundamentals of Queueing Theory , 2008 .

[34]  Vikram Krishnamurthy,et al.  ${Q}$-Learning Algorithms for Constrained Markov Decision Processes With Randomized Monotone Policies: Application to MIMO Transmission Control , 2007, IEEE Transactions on Signal Processing.

[35]  Anind K. Dey,et al.  Navigate like a cabbie: probabilistic reasoning from observed context-aware behavior , 2008, UbiComp.

[36]  Huseyin Topaloglu,et al.  Exploiting the Structural Properties of the Underlying Markov Decision Problem in the Q-Learning Algorithm , 2008, INFORMS J. Comput..

[37]  Emiliano Miluzzo,et al.  CenceMe - Injecting Sensing Presence into Social Networking Applications , 2007, EuroSSC.