Accoustate: Auto-annotation of IMU-generated Activity Signatures under Smart Infrastructure

—Human activities within smart infrastructures gen- erate a vast amount of IMU data from the wearables worn by individuals. Many existing studies rely on such sensory data for human activity recognition (HAR); however, one of the major bottlenecks is their reliance on pre-annotated or labeled data. Manual human-driven annotations are neither scalable nor efficient, whereas existing auto-annotation techniques heavily depend on video signatures. Still, video-based auto-annotation needs high computation resources and has privacy concerns when the data from a personal space, like a smart-home, is transferred to the cloud. This paper exploits the acoustic signatures generated from human activities to label the wearables’ IMU data at the edge, thus mitigating resource requirement and data privacy concerns. We utilize acoustic-based pre-trained HAR models for cross-modal labeling of the IMU data even when two individuals perform simultaneous but different activities under the same environmental context. We observe that non-overlapping acoustic gaps exist with a high probability during the simultaneous activities performed by two individuals in the environment’s acoustic context, which helps us resolve the overlapping activity signatures to label them individually. A principled evaluation of the proposed approach on two real-life in-house datasets further augmented to create a dual occupant setup, shows that the framework can correctly annotate a significant volume of unlabeled IMU data from both individuals with an accuracy of 82 . 59 % ( ± 17 . 94 % ) and 98 . 32 % ( ± 3 . 68 % ), respectively, for a workshop and a kitchen environment.

[1]  Shahriar Nirjon,et al.  Sound-Adapter: Multi-Source Domain Adaptation for Acoustic Classification Through Domain Discovery , 2021, IPSN.

[2]  Marcus Rohrbach,et al.  SMART Frame Selection for Action Recognition , 2020, AAAI.

[3]  Theodora Chaspari,et al.  Assessing ADL Routine Variability from High-dimensional Sensing Data using Hierarchical Clustering , 2020, BuildSys@SenSys.

[4]  Bivas Mitra,et al.  LASO: Exploiting Locomotive and Acoustic Signatures over the Edge to Annotate IMU Data for Human Activity Recognition , 2020, ICMI.

[5]  Jaime Teevan,et al.  Optimizing for Happiness and Productivity: Modeling Opportune Moments for Transitions and Breaks at Work , 2020, CHI.

[6]  Kun Wang,et al.  Sequential Weakly Labeled Multiactivity Localization and Recognition on Wearable Sensors Using Recurrent Attention Networks , 2020, IEEE Transactions on Human-Machine Systems.

[7]  Hae Young Noh,et al.  IDIoT: Towards Ubiquitous Identification of IoT Devices through Visual and Inertial Orientation Matching During Human Activity , 2020, 2020 IEEE/ACM Fifth International Conference on Internet-of-Things Design and Implementation (IoTDI).

[8]  Hae Young Noh,et al.  Fine-Grained Recognition of Activities of Daily Living through Structural Vibration and Electrical Sensing , 2019, BuildSys@SenSys.

[9]  Paul Lukowicz,et al.  Let there be IMU data: generating training data for wearable, motion sensor based activity recognition from monocular RGB videos , 2019, UbiComp/ISWC Adjunct.

[10]  Nirmalya Roy,et al.  Active Deep Learning for Activity Recognition with Context Aware Annotator Selection , 2019, KDD.

[11]  Gierad Laput,et al.  Sensing Fine-Grained Hand Activity with Smartwatches , 2019, CHI.

[12]  Samaneh Aminikhanghahi,et al.  Real-Time Change Point Detection with Application to Smart Home Time Series Data , 2019, IEEE Transactions on Knowledge and Data Engineering.

[13]  Fausto Giunchiglia,et al.  Fixing Mislabeling by Human Annotators Leveraging Conflict Resolution and Prior Knowledge , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[14]  Gierad Laput,et al.  Ubicoustics: Plug-and-Play Acoustic Activity Recognition , 2018, UIST.

[15]  Vaishali Sahu,et al.  IMU-Based Robust Human Activity Recognition using Feature Analysis, Extraction, and Reduction , 2018, 2018 24th International Conference on Pattern Recognition (ICPR).

[16]  Manmeet Mahinderjit Singh,et al.  Automatic Annotation of Unlabeled Data from Smartphone-Based Motion and Location Sensors , 2018, Sensors.

[17]  Josef Hallberg,et al.  Automatic Annotation for Human Activity Recognition in Free Living Using a Smartphone , 2018, Sensors.

[18]  Mani B. Srivastava,et al.  Enabling Edge Devices that Learn from Each Other: Cross Modal Training for Activity Recognition , 2018, EdgeSys@MobiSys.

[19]  Yaser Sheikh,et al.  Realtime Multi-person 2D Pose Estimation Using Part Affinity Fields , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Apostol Natsev,et al.  YouTube-8M: A Large-Scale Video Classification Benchmark , 2016, ArXiv.

[21]  Joan Claudi Socoró,et al.  A Review of Physical and Perceptual Feature Extraction Techniques for Speech, Music and Environmental Sounds , 2016 .

[22]  Nirmalya Roy,et al.  Active learning enabled activity recognition , 2016, 2016 IEEE International Conference on Pervasive Computing and Communications (PerCom).

[23]  Justin Salamon,et al.  A Dataset and Taxonomy for Urban Sound Research , 2014, ACM Multimedia.

[24]  Jun Li,et al.  Crowd++: unsupervised speaker count with smartphones , 2013, UbiComp.

[25]  DeLiang Wang,et al.  Analyzing noise robustness of MFCC and GFCC features in speaker identification , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[26]  Özlem Durmaz Incel,et al.  ARAS human activity datasets in multiple homes with multiple residents , 2013, 2013 7th International Conference on Pervasive Computing Technologies for Healthcare and Workshops.

[27]  Michel Vacher,et al.  SVM-Based Multimodal Classification of Activities of Daily Living in Health Smart Homes: Sensors, Algorithms, and First Experimental Results , 2010, IEEE Transactions on Information Technology in Biomedicine.

[28]  Martial Hebert,et al.  Temporal segmentation and activity classification from first-person sensing , 2009, 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[29]  Yoshinobu Kawahara,et al.  Change-Point Detection in Time-Series Data Based on Subspace Identification , 2007, Seventh IEEE International Conference on Data Mining (ICDM 2007).

[30]  David A. Cohn,et al.  Improving generalization with active learning , 1994, Machine Learning.

[31]  P. Rousseeuw Silhouettes: a graphical aid to the interpretation and validation of cluster analysis , 1987 .

[32]  M. Csíkszentmihályi,et al.  Validity and Reliability of the Experience‐Sampling Method , 1987, The Journal of nervous and mental disease.

[33]  Nirmalya Roy,et al.  A smart segmentation technique towards improved infrequent non-speech gestural activity recognition model , 2017, Pervasive Mob. Comput..

[34]  Frederic Ringsleben,et al.  Automated Annotation of Sensor data for Activity Recognition using Deep Learning , 2017, GI-Jahrestagung.

[35]  Oliver Kramer,et al.  K-Nearest Neighbors , 2013 .

[36]  Amy Loutfi,et al.  Automatic Annotation of Sensor Data Streams using Abductive Reasoning , 2013, KEOD.