| Literature DB >> 30231472 |
Qizhen Zhou1, Jianchun Xing2, Wei Chen3, Xuewei Zhang4, Qiliang Yang5.
Abstract
Gesture recognition acts as a key enabler for user-friendly human-computer interfaces (HCI). To bridge the human-computer barrier, numerous efforts have been devoted to designing accurate fine-grained gesture recognition systems. Recent advances in wireless sensing hold promise for a ubiquitous, non-invasive and low-cost system with existing Wi-Fi infrastructures. In this paper, we propose DeepNum, which enables fine-grained finger gesture recognition with only a pair of commercial Wi-Fi devices. The key insight of DeepNum is to incorporate the quintessence of deep learning-based image processing so as to better depict the influence induced by subtle finger movements. In particular, we make multiple efforts to transfer sensitive Channel State Information (CSI) into depth radio images, including antenna selection, gesture segmentation and image construction, followed by noisy image purification using high-dimensional relations. To fulfill the restrictive size requirements of deep learning model, we propose a novel region-selection method to constrain the image size and select qualified regions with dominant color and texture features. Finally, a 7-layer Convolutional Neural Network (CNN) and SoftMax function are adopted to achieve automatic feature extraction and accurate gesture classification. Experimental results demonstrate the excellent performance of DeepNum, which recognizes 10 finger gestures with overall accuracy of 98% in three typical indoor scenarios.Entities:
Keywords: channel state information; deep learning; gesture recognition; image processing
Mesh:
Year: 2018 PMID: 30231472 PMCID: PMC6165566 DOI: 10.3390/s18093142
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1The system architecture of DeepNum.
Figure 2Antenna Selection.
Figure 3Illustration of motion detection and segmentation.
Figure 4Image Construction.
Figure 5HOSVD-based Image De-noising.
Figure 6Experiment Study by Comparing Color and Texture Features in Different Regions.
Figure 7CNN Architecture.
Figure 8Experimental setup.
Figure 9Performance evaluation.
A comparison of state-of-the-art works for Wi-Fi-based gesture recognition.
| Properties | Year | Granularity | Signal | Segmentation? | Classification | Accuracy |
|---|---|---|---|---|---|---|
| 2015 | Hand | Amp | × | × | 91% | |
| 2015 | Finger | Amp | √ | DTW | 93.5% | |
| 2016 | Finger | Amp | √ | KNN + DTW | 90.4% | |
| 2016 | Finger | Amp | √ | DTW | 93% | |
| 2016 | Finger | Amp | × | DNN | 94% | |
| 2017 | Hand | Amp | √ | KNN | 91.4% | |
| Wang J. et al. [ | 2017 | Body | Amp + Phase | × | DNN | 90% |
| Wang X. et al. [ | 2017 | Body | Amp + AOA | × | CNN | 87% |
| 2018 | Finger | Amp + Phase | × | CNN | 98% | |
|
| 2018 | Finger | Amp + Phase |
| CNN | 98% |
Statistics of participants.
| User ID | Height/Weight (cm/kg) | BMI | User ID | Height/Weight (cm/kg) | BMI |
|---|---|---|---|---|---|
| 1 | 179/79 | 24.7 | 6 | 176/75 | 24.2 |
| 2 | 182/68 | 20.5 | 7 | 172/55 | 18.6 |
| 3 | 180/70 | 21.6 | 8 | 181/68 | 20.8 |
| 4 | 177/65 | 20.8 | 9 | 167/52 | 18.7 |
| 5 | 183/73 | 21.8 | 10 | 165/55 | 20.2 |
Figure 10Impact of user diversity.
Figure 11Impact of training strategy.
Figure 12Impact of raw material.
Figure 13Impact of image de-noising.
Figure 14Impact of antenna height.
Figure 15Impact of sampling rate.
Accuracy of DeepNum using different features.
| Experiments | Features | W/O | Color | Texture | Ours |
|---|---|---|---|---|---|
| Meeting Room | Accuracy (%) | 58.5 | 84.5 | 86.3 | 98.6 |
| Corridor | Accuracy (%) | 63.3 | 90.1 | 92.7 | 99.2 |
| Student Office | Accuracy (%) | 51.8 | 78.9 | 83.4 | 97.5 |
Running time of processing parts per gesture.
| Parts | Signal Processing | Image De-Noising | Region Selection | CNN Classification | Total |
|---|---|---|---|---|---|
| Times (ms) | 5.26 | 10.52 | 25.33 | 16.74 | 57.85 |