KU-HAR: An open dataset for heterogeneous human activity recognition

ABSTRACT In Artificial Intelligence, Human Activity Recognition (HAR) refers to the capability of machines to identify various activities performed by the users. The knowledge acquired from these recognition systems is integrated into many applications where the associated device uses it to identify actions or gestures and performs predefined tasks in response. HAR requires a large quantity of meticulously collected action data of diverse nature to fuel its learning algorithms. This paper aims to introduce a new set of HAR data collected from 90 participants who are 18 to 34 years old. The constructed dataset is named KU-HAR, and it contains 1945 raw activity samples that belong to 18 different classes. We used built-in smartphone sensors (accelerometer and gyroscope) to collect these HAR samples. Apart from the original (raw) time-domain samples, the dataset contains 20,750 subsamples (extracted from them) provided separately, each containing 3 seconds of data of the corresponding activity. Some classification results have been provided as well to propound the quality of the proposed dataset. The acquired results show that Random Forest (RF), an ensemble learning algorithm, can classify the subsamples with almost 90% accuracy. This dataset will enable smartphones and other smart devices to identify new activities and help researchers to design more delicate models based on practical HAR data.

[1]  Gary M. Weiss,et al.  Activity recognition using cell phone accelerometers , 2011, SKDD.

[2]  Allen Y. Yang,et al.  Distributed recognition of human actions using wearable motion sensor networks , 2009, J. Ambient Intell. Smart Environ..

[3]  Davide Anguita,et al.  Human Activity Recognition on Smartphones with Awareness of Basic Activities and Postural Transitions , 2014, ICANN.

[4]  Md. Atiqur Rahman Ahad,et al.  Feature Extraction, Performance Analysis and System Design Using the DU Mobility Dataset , 2018, IEEE Access.

[5]  Ricardo Chavarriaga,et al.  The Opportunity challenge: A benchmark database for on-body sensor-based activity recognition , 2013, Pattern Recognit. Lett..

[6]  Eduardo Casilari-Pérez,et al.  UMAFall: A Multisensor Dataset for the Research on Automatic Fall Detection , 2017, FNC/MobiSPC.

[7]  Nobuhiko Nishio,et al.  HASC Challenge: gathering large scale human activity corpus for the real-world activity understandings , 2011, AH '11.

[8]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[9]  Md. Atiqur Rahman Ahad,et al.  DU-MD: An Open-Source Human Action Dataset for Ubiquitous Wearable Sensors , 2018, 2018 Joint 7th International Conference on Informatics, Electronics & Vision (ICIEV) and 2018 2nd International Conference on Imaging, Vision & Pattern Recognition (icIVPR).

[10]  Miguel A. Labrador,et al.  Human Activity Recognition: Using Wearable Sensors and Smartphones , 2013 .

[11]  Enrique Fernández-Blanco,et al.  A Public Domain Dataset for Real-Life Human Activity Recognition Using Smartphone Sensors , 2020, Sensors.

[12]  Juan Carlos Fernández,et al.  Multiobjective evolutionary algorithms to identify highly autocorrelated areas: the case of spatial distribution in financially compromised farms , 2014, Ann. Oper. Res..

[13]  Niloy Sikder,et al.  Human Activity Recognition Using Multichannel Convolutional Neural Network , 2019, 2019 5th International Conference on Advances in Electrical Engineering (ICAEE).

[14]  Davide Anguita,et al.  A Public Domain Dataset for Human Activity Recognition using Smartphones , 2013, ESANN.

[15]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[16]  Mi Zhang,et al.  USC-HAD: a daily activity dataset for ubiquitous activity recognition using wearable sensors , 2012, UbiComp.