Robust Unsupervised Factory Activity Recognition with Body-worn Accelerometer Using Temporal Structure of Multiple Sensor Data Motifs

This paper presents a robust unsupervised method for recognizing factory work using sensor data from body-worn acceleration sensors. In line-production systems, each factory worker repetitively performs a predefined work process with each process consisting of a sequence of operations. Because of the difficulty in collecting labeled sensor data from each factory worker, unsupervised factory activity recognition has been attracting attention in the ubicomp community. However, prior unsupervised factory activity recognition methods can be adversely affected by any outlier activities performed by the workers. In this study, we propose a robust factory activity recognition method that tracks frequent sensor data motifs, which can correspond to particular actions performed by the workers, that appear in each iteration of the work processes. Specifically, this study proposes tracking two types of motifs: period motifs and action motifs, during the unsupervised recognition process. A period motif is a unique data segment that occurs only once in each work period (one iteration of an overall work process). An action motif is a data segment that occurs several times in each work period, corresponding to an action that is performed several times in each period. Tracking multiple period motifs enables us to roughly capture the temporal structure and duration of the work period even when outlier activities occur. Action motifs, which are spread throughout the work period, permit us to precisely detect the start time of each operation. We evaluated the proposed method using sensor data collected from workers in actual factories and achieved state-of-the-art performance.

[1]  Kristof Van Laerhoven,et al.  Detecting leisure activities with dense motif discovery , 2012, UbiComp.

[2]  Franco Zambonelli,et al.  XLearn: Learning Activity Labels across Heterogeneous Datasets , 2020, ACM Trans. Intell. Syst. Technol..

[3]  T. Kirste,et al.  Computational State Space Models for Activity and Intention Recognition. A Feasibility Study , 2014, PloS one.

[4]  Marcel Bogers,et al.  on Intelligent Manufacturing and Automation , 2013 The Smart Factory : Exploring Adaptive and Flexible Manufacturing Solutions , 2014 .

[5]  Gerhard Tröster,et al.  Fusion of String-Matched Templates for Continuous Activity Recognition , 2007, 2007 11th IEEE International Symposium on Wearable Computers.

[6]  Matthew J. Johnson,et al.  Bayesian nonparametric hidden semi-Markov models , 2012, J. Mach. Learn. Res..

[7]  Juhi Ranjan,et al.  Object hallmarks: identifying object users using wearable wrist sensors , 2015, UbiComp.

[8]  Eamonn J. Keogh,et al.  Dimensionality Reduction for Fast Similarity Search in Large Time Series Databases , 2001, Knowledge and Information Systems.

[9]  Mario Aehnelt,et al.  Learn but work!: towards self-directed learning at mobile assembly workplaces , 2015, I-KNOW.

[10]  Thomas Kirste,et al.  Computational causal behaviour models for assisted manufacturing , 2015, iWOAR.

[11]  Gregory D. Abowd,et al.  A practical approach for recognizing eating moments with wrist-mounted inertial sensing , 2015, UbiComp.

[12]  Alex Pentland,et al.  InSense: Interest-Based Life Logging , 2006, IEEE MultiMedia.

[13]  Kristina Yordanova,et al.  Extracting Planning Operators from Instructional Texts for Behaviour Interpretation , 2018, KI.

[14]  Albrecht Schmidt,et al.  The potentials of in-situ-projection for augmented workplaces in production: a study with impaired persons , 2013, CHI Extended Abstracts.

[15]  Sebastian Bader,et al.  Information Assistance for Smart Assembly Stations , 2015, ICAART.

[16]  Gernot A. Fink,et al.  Convolutional Neural Networks for Human Activity Recognition Using Body-Worn Sensors , 2018, Informatics.

[17]  Juan Ye,et al.  Unsupervised domain adaptation for activity recognition across heterogeneous datasets , 2020, Pervasive Mob. Comput..

[18]  Paul Lukowicz,et al.  Combining Motion Sensors and Ultrasonic Hands Tracking for Continuous Activity Recognition in a Maintenance Scenario , 2006, 2006 10th IEEE International Symposium on Wearable Computers.

[19]  Takuya Maekawa,et al.  Unsupervised Factory Activity Recognition with Wearable Sensors Using Process Instruction Information , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[20]  Paul Lukowicz,et al.  WearNET: A Distributed Multi-sensor System for Context Aware Wearables , 2002, UbiComp.

[21]  Gaetano Borriello,et al.  A Practical Approach to Recognizing Physical Activities , 2006, Pervasive.

[22]  Takuya Maekawa,et al.  Toward practical factory activity recognition: unsupervised understanding of repetitive assembly work in a factory , 2016, UbiComp.

[23]  Carmen Constantinescu,et al.  Smart Factory - A Step towards the Next Generation of Manufacturing , 2008 .

[24]  Takuya Maekawa,et al.  An energy-aware method for the joint recognition of activities and gestures using wearable sensors , 2015, SEMWEB.

[25]  Bernt Schiele,et al.  Discovery of activity patterns using topic models , 2008 .

[26]  Jessica Lin,et al.  Finding Motifs in Time Series , 2002, KDD 2002.

[27]  Irfan A. Essa,et al.  Discovering Characteristic Actions from On-Body Sensor Data , 2006, 2006 10th IEEE International Symposium on Wearable Computers.

[28]  Daniel Roggen,et al.  Deep convolutional feature transfer across mobile activity recognition domains, sensor modalities and locations , 2016, SEMWEB.

[29]  Takuya Maekawa,et al.  WristSense: Wrist-worn sensor device with camera for daily activity recognition , 2012, 2012 IEEE International Conference on Pervasive Computing and Communications Workshops.

[30]  Takuya Maekawa,et al.  Recognizing Handheld Electrical Device Usage with Hand-Worn Coil of Wire , 2012, Pervasive.

[31]  Takuya Maekawa,et al.  Unsupervised Activity Recognition with User's Physical Characteristics Data , 2011, 2011 15th Annual International Symposium on Wearable Computers.

[32]  S. Chiba,et al.  Dynamic programming algorithm optimization for spoken word recognition , 1978 .

[33]  Pekka Siirtola,et al.  Activity recognition using a wrist-worn inertial measurement unit: A case study for industrial assembly lines , 2009, 2009 17th Mediterranean Conference on Control and Automation.

[34]  Hans-Peter Wiendahl,et al.  Montage in der industriellen Produktion : ein Handbuch für die Praxis , 2012 .

[35]  Takuya Maekawa,et al.  Selecting home appliances with smart glass based on contextual information , 2016, UbiComp.

[36]  Bodo Urban,et al.  Using Activity Recognition for the Tracking of Assembly Processes : Challenges and Requirements , 2014 .

[37]  Bodo Urban,et al.  The Knowledge Gap: Providing Situation-Aware Information Assistance on the Shop Floor , 2015, HCI.

[38]  Paul Lukowicz,et al.  Gesture spotting using wrist worn microphone and 3-axis accelerometer , 2005, sOc-EUSAI '05.

[39]  Paul Lukowicz,et al.  Recognizing Workshop Activity Using Body Worn Microphones and Accelerometers , 2004, Pervasive.