Towards Domain-independent Complex and Fine-grained Gesture Recognition with RFID

Gesture recognition plays a fundamental role in emerging Human-Computer Interaction (HCI) paradigms. Recent advances in wireless sensing show promise for device-free and pervasive gesture recognition. Among them, RFID has gained much attention given its low-cost, light-weight and pervasiveness, but pioneer studies on RFID sensing still suffer two major problems when it comes to gesture recognition. The first is they are only evaluated on simple whole-body activities, rather than complex and fine-grained hand gestures. The second is they can not effectively work without retraining in new domains, i.e. new users or environments. To tackle these problems, in this paper, we propose RFree-GR, a domain-independent RFID system for complex and fine-grained gesture recognition. First of all, we exploit signals from the multi-tag array to profile the sophisticated spatio-temporal changes of hand gestures. Then, we elaborate a Multimodal Convolutional Neural Network (MCNN) to aggregate information across signals and abstract complex spatio-temporal patterns. Furthermore, we introduce an adversarial model to our deep learning architecture to remove domain-specific information while retaining information relevant to gesture recognition. We extensively evaluate RFree-GR on 16 commonly used American Sign Language (ASL) words. The average accuracy for new users and environments (new setup and new position) are $89.03%$, $90.21%$ and $88.38%$, respectively, significantly outperforming existing RFID based solutions, which demonstrates the superior effectiveness and generalizability of RFree-GR.

[1]  Jian Liu,et al.  Multi - Touch in the Air: Device-Free Finger Tracking and Gesture Recognition via COTS RFID , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.

[2]  Wei Xi,et al.  RFIPad: Enabling Cost-Efficient and Device-Free In-air Handwriting Using Passive Tags , 2017, 2017 IEEE 37th International Conference on Distributed Computing Systems (ICDCS).

[3]  Dong Wang,et al.  Au-Id , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[4]  Parth H. Pathak,et al.  mmASL: Environment-Independent ASL Gesture Recognition Using 60 GHz Millimeter-wave Signals , 2020, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[5]  Chris Harrison,et al.  BeamBand: Hand Gesture Sensing with Ultrasonic Beamforming , 2019, CHI.

[6]  Panlong Yang,et al.  SignSpeaker: A Real-time, High-Precision SmartWatch-based Sign Language Translator , 2019, MobiCom.

[7]  Jie Yang,et al.  E-eyes: device-free location-oriented activity identification using fine-grained WiFi signatures , 2014, MobiCom.

[8]  Umberto Spagnolini,et al.  Device-Free Radio Vision for Assisted Living: Leveraging wireless channel quality information for human sensing , 2016, IEEE Signal Processing Magazine.

[9]  Kaishun Wu,et al.  GRfid: A Device-Free RFID-Based Gesture Recognition System , 2017, IEEE Transactions on Mobile Computing.

[10]  Gierad Laput,et al.  ViBand: High-Fidelity Bio-Acoustic Sensing Using Commodity Smartwatch Accelerometers , 2016, UIST.

[11]  Sanjeev Khudanpur,et al.  Audio augmentation for speech recognition , 2015, INTERSPEECH.

[12]  Hagen Soltau,et al.  Neural Speech Recognizer: Acoustic-to-Word LSTM Model for Large Vocabulary Speech Recognition , 2016, INTERSPEECH.

[13]  Muhammad Shahzad,et al.  Multi-User Gesture Recognition Using WiFi , 2018, MobiSys.

[14]  Lei Yang,et al.  Tagoram: real-time tracking of mobile RFID tags to high precision using COTS devices , 2014, MobiCom.

[15]  Qian Zhang,et al.  AdaRF: Adaptive RFID-based Indoor Localization Using Deep Learning Enhanced Holography , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[16]  Song Guo,et al.  ReActor: Real-time and Accurate Contactless Gesture Recognition with RFID , 2019, 2019 16th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON).

[17]  Bo Chen,et al.  RFree-ID: An Unobtrusive Human Identification System Irrespective of Walking Cofactors Using COTS RFID , 2018, 2018 IEEE International Conference on Pervasive Computing and Communications (PerCom).

[18]  Pavlo Molchanov,et al.  Online Detection and Classification of Dynamic Hand Gestures with Recurrent 3D Convolutional Neural Networks , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  Ivan Poupyrev,et al.  Soli , 2016, ACM Trans. Graph..

[20]  Shuangquan Wang,et al.  SignFi , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[21]  He Wang,et al.  I am a Smartwatch and I can Track my User's Arm , 2016, MobiSys.

[22]  Dina Katabi,et al.  RF-IDraw: virtual touch screen in the air using RF signals , 2014, S3 '14.

[23]  Gierad Laput,et al.  Sensing Fine-Grained Hand Activity with Smartwatches , 2019, CHI.

[24]  Tommi S. Jaakkola,et al.  Learning Sleep Stages from Radio Signals: A Conditional Adversarial Architecture , 2017, ICML.

[25]  Jake Araullo,et al.  The Leap Motion controller: a view on sign language , 2013, OZCHI.

[26]  Dan Wu,et al.  FullBreathe , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[27]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[28]  Lei Yang,et al.  See Through Walls with COTS RFID System! , 2015, MobiCom.

[29]  Mi Zhang,et al.  DeepASL: Enabling Ubiquitous and Non-Intrusive Word and Sentence-Level Sign Language Translation , 2017, SenSys.

[30]  Zhengyou Zhang,et al.  Microsoft Kinect Sensor and Its Effect , 2012, IEEE Multim..

[31]  Jiangchuan Liu,et al.  TagFree Activity Identification with RFIDs , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[32]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[33]  Ivan Poupyrev,et al.  Interacting with Soli: Exploring Fine-Grained Dynamic Gesture Recognition in the Radio-Frequency Spectrum , 2016, UIST.

[34]  Victor S. Lempitsky,et al.  Unsupervised Domain Adaptation by Backpropagation , 2014, ICML.

[35]  Bingbing Ni,et al.  Recurrent Modeling of Interaction Context for Collective Activity Recognition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[36]  Chenglin Miao,et al.  Towards Environment Independent Device Free Human Activity Recognition , 2018, MobiCom.

[37]  Dong Wang,et al.  CRH: A Contactless Respiration and Heartbeat Monitoring System with COTS RFID Tags , 2018, 2018 15th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON).

[38]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[39]  Daniel Vogel,et al.  WRIST: Watch-Ring Interaction and Sensing Technique for Wrist Gestures and Macro-Micro Pointing , 2019, MobileHCI.

[40]  Wei Wang,et al.  Device-free gesture tracking using acoustic signals , 2016, MobiCom.

[41]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[42]  Lihua Xie,et al.  Learning Gestures From WiFi: A Siamese Recurrent Convolutional Architecture , 2019, IEEE Internet of Things Journal.

[43]  Qian Zhang,et al.  MyoSign: enabling end-to-end sign language recognition with wearables , 2019, IUI.

[44]  Chin-Hui Lee,et al.  Sign Transition Modeling and a Scalable Solution to Continuous Sign Language Recognition for Real-World Applications , 2016, ACM Trans. Access. Comput..

[45]  Sheng Tan,et al.  WiFinger: leveraging commodity WiFi for fine-grained finger gesture recognition , 2016, MobiHoc.