Please wait a minute...
Journal of ZheJiang University (Engineering Science)  2023, Vol. 57 Issue (3): 503-511    DOI: 10.3785/j.issn.1008-973X.2023.03.008
    
Hand gesture/state recognition based on inertial measurement unit at high sample rate
Zhuo-feng LI(),Ming-hui SUN*()
Department of Computer Science and Technology, Jilin University, Changchun 130012, China
Download: HTML     PDF(3956KB) HTML
Export: BibTeX | EndNote (RIS)      

Abstract  

In order to realize gesture recognition and hand state recognition at the same time, a single inertial measurement unit-based gesture recognition and touch recognition prototype was built, considering the inertial measurement unit at high sample rate has the capability of collecting motion signals and vibration signals simultaneously. The differences within hand state data and gesture data in the time and frequency domains were visually analyzed. Hand state, slipping gesture and circling gesture data sets were established. Considering the difference within data features, differential feature extraction methods were proposed, and neural network structures for hand state classification and gesture classification were constructed. Neural network models were trained by the data sets to achieve 99% accuracy rate in the comprehensive hand state recognition task, and 98% accuracy rate in both the slipping gesture recognition task and the circling gesture recognition task. A prototype program framework for real-time data stream processing, state shifting, and unknown class judgment was proposed. And a real-time program based on the hand state recognition model entities and the gesture recognition model entities was built, and the overall computational latency of the actual operation and the single model computational latency were measured, in order to prove the capability of real-time computing. Experimental results of model evaluation and real-time computing verification showed that, accurate and real-time hand states and gesture recognition with high sample rate inertial measurement units was feasible.



Key wordshuman-computer interaction      inertial measurement unit      hand gesture recognition      touch recognition      wearable device     
Received: 12 March 2022      Published: 31 March 2023
CLC:  TP 334.2  
  TP 183  
Fund:  国家自然科学基金资助项目(61872164)
Corresponding Authors: Ming-hui SUN     E-mail: zfli20@mails.jlu.edu.cn;smh@jlu.edu.cn
Cite this article:

Zhuo-feng LI,Ming-hui SUN. Hand gesture/state recognition based on inertial measurement unit at high sample rate. Journal of ZheJiang University (Engineering Science), 2023, 57(3): 503-511.

URL:

https://www.zjujournals.com/eng/10.3785/j.issn.1008-973X.2023.03.008     OR     https://www.zjujournals.com/eng/Y2023/V57/I3/503


基于高采样率惯性测量单元的手部状态与手势识别

为了同时实现手势识别与手部状态识别,针对高采样率惯性测量单元具有同时采集动作信号与振动信号的特点,搭建基于单惯性测量单元的手势识别与触摸识别原型设备. 可视化分析手部状态数据与手势数据在时域与频域上的差异,建立手部状态、划动手势与画圈手势数据集. 针对数据特征的差异,提出差异化特征提取方法,分别构建手部状态分类与手势分类的神经网络结构. 使用数据集训练神经网络模型,在手部综合状态识别任务中正确率达到99%,在划动手势识别任务和画圈手势识别任务中的正确率均达到98%. 提出实时数据流处理、状态转移、未知类别判断的原型程序框架,基于手部状态识别模型实体与手势识别模型实体搭建实时程序,测量实际运行整体计算延时与单模型计算延时,验证模型实时运算能力. 模型评估实验与实时运算能力验证实验结果表明,使用高采样率惯性测量单元准确且实时地识别手部状态与手势具备可行性.


关键词: 人机交互,  惯性测量单元,  手势识别,  触摸识别,  可穿戴设备 
Fig.1 Example of hardware wearing
Fig.2 Example of hardware connection
Fig.3 Raw signals, low-frequency signals and high-frequency signals of single-axis acceleration in time domain
Fig.4 Frequency band energy distribution for different scenarios
Fig.5 Visualization of t-SNE for each classification data on different features
模型名称 分类数量 模型分类类别
A 2 未佩戴、佩戴时手部静止
B 2 划动、画圈
C 2 顺时针、逆时针
D 3 静止(包括未佩戴和手部静止)、点击、
手部移动(包括划动与画圈与空中运动)
E 3 (划动)左、居中、右
F 3 (划动)上、居中、下
G 2 表面上移动、非表面上移动
Tab.1 Task decomposition for models
Fig.6 Structure of feature classifier
Fig.7 Structure of autoencoder
模型名称 N A/% P/% R/% F/%
A 240 97.92
(0.92)
100.00
(0.00)
95.83
(1.84)
97.86
(0.95)
B 960 90.63
(0.78)
90.63
(1.76)
90.67
(0.95)
90.63
(0.69)
C 620 98.63
(1.32)
99.29
(1.61)
97.97
(1.03)
98.62
(1.31)
D 600 99.22
(0.14)
99.30
(0.22)
98.75
(0.19)
99.01
(0.17)
E 480 99.77
(0.18)
99.73
(0.23)
99.76
(0.19)
99.75
(0.20)
F 480 94.23
(3.51)
93.94
(3.09)
94.64
(3.18)
93.77
(3.72)
G 1920 98.42
(0.23)
99.21
(0.37)
97.63
(0.53)
98.41
(0.23)
Tab.2 Means and standard deviations of classification models evaluation indices
Fig.8 Confusion matrix for two classification tasks
Fig.9 State diagram of real-time program
Fig.10 Discriminant tree for unknown class
模型名称 T/ms 模型名称 T/ms
A 0.92(1.95) E 1.28(1.86)
C 1.04(1.34) F 2.02(3.28)
D 0.99(1.75) G 1.15(2.76)
Tab.3 Means and standard deviations of model calculation time
[1]   SIEK K A, ROGERS Y, CONNELLY K H. Fat finger worries: how older and younger users physically interact with PDAs [C]// IFIP Conference on Human-Computer Interaction. [S.l.]: Springer, 2005: 267-280.
[2]   ZHOU J, ZHANG Y, LAPUT G, et al. AuraSense: enabling expressive around-smartwatch interactions with electric field sensing [C]// Proceedings of the 29th Annual Symposium on User Interface Software and Technology. [S.l.]: Association for Computing Machinery, 2016: 81-86.
[3]   ZHANG Y, KIENZLE W, MA Y, et al. ActiTouch: robust touch detection for on-skin AR/VR interfaces [C]// Proceedings of the 32nd Annual ACM Symposium on User Interface Software and Technology. [S.l.]: Association for Computing Machinery, 2019: 1151-1159.
[4]   LAPUT G, HARRISON C. SurfaceSight: a new spin on touch, user, and object sensing for IoT experiences [C]// Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. [S.l.]: Association for Computing Machinery, 2019: 1-12.
[5]   SUKKARIEH S, NEBOT E M, DURRANT-WHYTE H F A high integrity IMU/GPS navigation loop for autonomous land vehicle applications[J]. IEEE Transactions on Robotics and Automation, 1999, 15 (3): 572- 578
doi: 10.1109/70.768189
[6]   BU Y, XIE L, YIN Y, et al. Handwriting-assistant: reconstructing continuous strokes with millimeter-level accuracy via attachable inertial sensors [C]// Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies. [S.l.]: Association for Computing Machinery, 2021, 5(4): 1-25.
[7]   BAO L, INTILLE S S. Activity recognition from user-annotated acceleration data [C]// International Conference on Pervasive Computing. [S.l.]: Springer, 2004: 1-17.
[8]   GU Y, YU C, LI Z, et al. Accurate and low-latency sensing of touch contact on any surface with finger-worn IMU sensor [C]// Proceedings of the 32nd Annual ACM Symposium on User Interface Software and Technology. [S.l.]: Association for Computing Machinery, 2019: 1059-1070.
[9]   XU X, LI J, YUAN T, et al. HulaMove: using commodity IMU for waist interaction [C]// Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. [S.l.]: Association for Computing Machinery, 2021: 1-16.
[10]   OTT F, WEHBI M, HAMANN T, et al. The OnHW dataset: online handwriting recognition from IMU-enhanced ballpoint pens with machine learning [C]// Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies. [S.l.]: Association for Computing Machinery, 2020, 4(3): 1-20.
[11]   WEHBI M, HAMANN T, BARTH J, et al. Towards an IMU-based pen online handwriting recognizer [C]// International Conference on Document Analysis and Recognition. [S.l.]: Springer, 2021: 289-303.
[12]   XU X, GONG J, BRUM C, et al. Enabling hand gesture customization on wrist-worn devices [C]// Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems. [S.l.]: Springer, 2022: 1-19.
[13]   MEIER M, STRELI P, FENDER A, et al. TaplD: rapid touch interaction in virtual reality using wearable sensing [C]// 2021 IEEE Virtual Reality and 3D User Interfaces (VR). Lisboa: IEEE, 2021: 519-528.
[14]   LAPUT G, XIAO R, HARRISON C. Viband: high-fidelity bio-acoustic sensing using commodity smartwatch accelerometers [C]// Proceedings of the 29th Annual Symposium on User Interface Software and Technology. [S.l.]: Association for Computing Machinery, 2016: 321-333.
[15]   LAPUT G, HARRISON C. Sensing fine-grained hand activity with smartwatches [C]// Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. [S.l.]: Association for Computing Machinery, 2019: 1-13.
[16]   GONG J, GUPTA A, BENKO H. Acustico: surface tap detection and localization using wrist-based acoustic TDOA sensing [C]// Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology. [S.l.]: Association for Computing Machinery, 2020: 406-419.
[17]   MADGWICK S O H, HARRISON A J L, VAIDYANATHAN R. Estimation of IMU and MARG orientation using a gradient descent algorithm [C]// 2011 IEEE International Conference on Rehabilitation Robotics. Zurich: IEEE, 2011: 1-7.
[1] Shi-wei CHENG,Wei GUO. Reading annotation generation method through analysis of visual behavior and text features[J]. Journal of ZheJiang University (Engineering Science), 2020, 54(6): 1115-1125.
[2] Fei FEI,Shen-yu LIU,Chang-cheng WU,De-hua YANG,Sheng-li ZHOU. Human kinetic energy harvesting technology based on magnetic levitation structure[J]. Journal of ZheJiang University (Engineering Science), 2019, 53(11): 2215-2222.
[3] HU Li-sha, WANG Su-zhen, CHEN Yi-qiang, GAO Chen-long, HU Chun-yu, JIANG Xin-long, CHEN Zhen-yu, GAO Xing-yu. Fall detection algorithms based on wearable device: a review[J]. Journal of ZheJiang University (Engineering Science), 2018, 52(9): 1717-1728.
[4] JIANG Zhuo-hua, JIANG Huan-yu, TONG Jun-hua. Optimal design of end-effector on automatic plug seedling transplanter[J]. Journal of ZheJiang University (Engineering Science), 2017, 51(6): 1119-1125.
[5] GONG Yong, ZHANG San yuan, LIU Zhi fang3, SHEN Fa. Eye movement study on color effects to icon visual search efficiency[J]. Journal of ZheJiang University (Engineering Science), 2016, 50(10): 1987-1994.
[6] LU Dan-hui, ZHOU Wen-hui, GONG Xiao-jin, LIU Ji-lin. Decoupled mobile robot motion estimation based on fusion of
visual and inertial measurement unit
[J]. Journal of ZheJiang University (Engineering Science), 2012, 46(6): 1021-1026.
[7] CHEN Wei-dong,LI Xin,LIU Jun,HAO Yao-yao,LIAO Yu-xi,SU Yu,ZHANG Shao-min,ZHENG X. Mathematical morphology based electro-oculography recognition
algorithm for human-computer interaction
[J]. Journal of ZheJiang University (Engineering Science), 2011, 45(4): 644-649.