Summary of the Sussex-Huawei locomotion-transportation recognition challenge 2019

In this paper we summarize the contributions of participants to the Sussex-Huawei Transportation-Locomotion (SHL) Recognition Challenge organized at the HASCA Workshop of Ubi-Comp 2019. The goal of this machine learning/data science challenge is to recognize eight locomotion and transportation activities (Still, Walk, Run, Bike, Bus, Car, Train, Subway) from the inertial sensor data of a smartphone in a placement independent manner. The training data is collected with smartphones placed at three body positions (Torso, Bag and Hips), while the testing data is collected with a smartphone placed at another body position (Hand). We introduce the dataset used in the challenge and the protocol for the competition. We present a meta-analysis of the contributions from 14 submissions, their approaches, the software tools used, computational cost and the achieved results. Overall, three submissions achieved F1 scores between 70% and 80% five with F1 scores between 60% and 70%, five between between 50% and 60%, and one below 50%, with a latency of a maximum of 5 seconds.

[1]  Jani Bizjak,et al.  Winning the Sussex-Huawei Locomotion-Transportation Recognition Challenge , 2019, Human Activity Sensing.

[2]  Stefan Valentin,et al.  Enabling Reproducible Research in Sensor-Based Transportation Mode Recognition With the Sussex-Huawei Dataset , 2019, IEEE Access.

[3]  Benjamin Cauchi,et al.  Transportation mode classification from smartphone sensors via a long-short-term-memory network , 2019, UbiComp/ISWC Adjunct.

[4]  Sozo Inoue,et al.  Position independent activity recognition using shallow neural architecture and empirical modeling , 2019, UbiComp/ISWC Adjunct.

[5]  Longfei Zheng,et al.  Application of IndRNN for human activity recognition: the Sussex-Huawei locomotion-transportation challenge , 2019, UbiComp/ISWC Adjunct.

[6]  Haiyong Luo,et al.  Transportation recognition with the Sussex-Huawei Locomotion challenge , 2019, UbiComp/ISWC Adjunct.

[7]  Rahul C. Shah,et al.  Locomotion recognition using XGBoost and neural network ensemble , 2019, UbiComp/ISWC Adjunct.

[8]  Lin Wang,et al.  Summary of the Sussex-Huawei Locomotion-Transportation Recognition Challenge , 2018, UbiComp/ISWC Adjunct.

[9]  Peter Widhalm,et al.  Ensemble-based domain adaptation for transport mode recognition with mobile sensors , 2019, UbiComp/ISWC Adjunct.

[10]  Lin Wang,et al.  Benchmarking the SHL Recognition Challenge with Classical and Deep-Learning Pipelines , 2018, UbiComp/ISWC Adjunct.

[11]  Hristijan Gjoreski,et al.  Benchmark Performance for the Sussex-Huawei Locomotion and Transportation Recognition Challenge 2018 , 2019, Human Activity Sensing.

[12]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[13]  Lin Wang,et al.  Sound-based Transportation Mode Recognition with Smartphones , 2019, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[14]  Masaki Shuzo,et al.  Application of CNN for Human Activity Recognition with FFT Spectrogram of Acceleration and Gyro Sensors , 2018, UbiComp/ISWC Adjunct.

[15]  Dmitrijs Balabka,et al.  Semi-supervised learning for human activity recognition using adversarial autoencoders , 2019, UbiComp/ISWC Adjunct.

[16]  Daniel Roggen,et al.  Human and Machine Recognition of Transportation Modes from Body-Worn Camera Images , 2019, 2019 Joint 8th International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition (icIVPR).

[17]  Lin Wang,et al.  The University of Sussex-Huawei Locomotion and Transportation Dataset for Multimodal Analytics With Mobile Devices , 2018, IEEE Access.

[18]  Chih-Jen Lin,et al.  Big Data Small Footprint: The Design of A Low-Power Classifier for Detecting Transportation Modes , 2014, Proc. VLDB Endow..

[19]  Fang Zhao,et al.  Applying 1D sensor DenseNet to Sussex-Huawei locomotion-transportation recognition challenge , 2019, UbiComp/ISWC Adjunct.

[20]  Jun-Ho Choi,et al.  EmbraceNet for activity: a deep multimodal fusion architecture for activity recognition , 2019, UbiComp/ISWC Adjunct.

[21]  Masaki Shuzo,et al.  CNN for human activity recognition on small datasets of acceleration and gyro sensors using transfer learning , 2019, UbiComp/ISWC Adjunct.

[22]  Masud Ahmed,et al.  POIDEN: position and orientation independent deep ensemble network for the classification of locomotion and transportation modes , 2019, UbiComp/ISWC Adjunct.

[23]  Matjaz Gams,et al.  Cross-location transfer learning for the sussex-huawei locomotion recognition challenge , 2019, UbiComp/ISWC Adjunct.

[24]  Hao Xia,et al.  Using Smart Phone Sensors to Detect Transportation Modes , 2014, Sensors.