| Literature DB >> 31487894 |
Minwoo Kim1, Jaechan Cho2, Seongjoo Lee3, Yunho Jung4.
Abstract
We propose an efficient hand gesture recognition (HGR) algorithm, which can cope with time-dependent data from an inertial measurement unit (IMU) sensor and support real-time learning for various human-machine interface (HMI) applications. Although the data extracted from IMU sensors are time-dependent, most existing HGR algorithms do not consider this characteristic, which results in the degradation of recognition performance. Because the dynamic time warping (DTW) technique considers the time-dependent characteristic of IMU sensor data, the recognition performance of DTW-based algorithms is better than that of others. However, the DTW technique requires a very complex learning algorithm, which makes it difficult to support real-time learning. To solve this issue, the proposed HGR algorithm is based on a restricted column energy (RCE) neural network, which has a very simple learning scheme in which neurons are activated when necessary. By replacing the metric calculation of the RCE neural network with DTW distance, the proposed algorithm exhibits superior recognition performance for time-dependent sensor data while supporting real-time learning. Our verification results on a field-programmable gate array (FPGA)-based test platform show that the proposed HGR algorithm can achieve a recognition accuracy of 98.6% and supports real-time learning and recognition at an operating frequency of 150 MHz.Entities:
Keywords: dynamic time warping (DTW); hand gesture recognition (HGR); inertial measurement unit (IMU); machine learning; real-time learning; restricted coulomb energy (RCE) neural network
Mesh:
Year: 2019 PMID: 31487894 PMCID: PMC6767360 DOI: 10.3390/s19183827
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Structure of an RCE neural network.
Figure 2Time alignment of two data sequences; the aligned points are denoted by the arrows.
Figure 3Cost matrix.
Figure 4Accumulated cost matrix.
Figure 5Examples of the data sequences: (a) two sequences for label 0; (b) two sequences for label 7.
Figure 6Structure of the proposed HGR algorithm.
Figure 7HGR test platform: (a) photograph of the test platform; (b) configuration of the test platform.
Figure 8Histogram of the dataset: (a) label 0; (b) label 1; (c) label 2; (d) label 3; (e) label 4; (f) label 5; (g) label 6; (h) label 7; (i) label 8; (j) label 9.
Figure 9Block diagram of the proposed hand gesture recognizer.
Implementation results of the proposed hand gesture recognizer.
| Block | Neural Network | NCU | ANDU | Total |
|---|---|---|---|---|
| FPGA Logic Elements (/114,480) | 25,765 | 2540 | 2970 | 31,275 (27.32%) |
| Memory [bists] (/3,981,312) | 280,896 | 0 | 0 | 280,896 (7.06%) |
Figure 103D number dataset.
Confusion matrix of the proposed algorithm.
| Answer | Prediction | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | |
| 0 | 96% | 0% | 3% | 0% | 0% | 0% | 0% | 0% | 1% | 0% |
| 1 | 0% | 100% | 0% | 0% | 0% | 0% | 0% | 1% | 0% | 0% |
| 2 | 0% | 0% | 97% | 0% | 0% | 0% | 0% | 0% | 0% | 0% |
| 3 | 0% | 0% | 0% | 99% | 0% | 1% | 0% | 0% | 0% | 0% |
| 4 | 0% | 0% | 0% | 0% | 100% | 0% | 0% | 0% | 0% | 0% |
| 5 | 0% | 0% | 0% | 0% | 0% | 97% | 0% | 0% | 0% | 0% |
| 6 | 2% | 0% | 0% | 1% | 0% | 2% | 100% | 0% | 0% | 0% |
| 7 | 2% | 0% | 0% | 0% | 0% | 0% | 0% | 99% | 0% | 0% |
| 8 | 0% | 0% | 0% | 0% | 0% | 1% | 0% | 0% | 99% | 1% |
| 9 | 0% | 0% | 0% | 0% | 0% | 1% | 0% | 0% | 0% | 99% |
|
| 100% | 100% | 100% | 100% | 100% | 100% | 100% | 100% | 100% | 100% |
Recognition performance of the proposed algorithm and others.
| Algorithm | User1 | User2 | User3 | User4 | User5 | Average |
|---|---|---|---|---|---|---|
| RCE neural network | 82.5% | 88.0% | 81.0% | 92.5% | 83.0% | 85.4% |
| MLP | 81.5% | 91.5% | 86.5% | 91.0% | 89.5% | 88.0% |
| DTW-based HGR | 94.6% | 94.6% | 94.6% | 94.6% | 94.6% | 94.6% |
| Proposed | 99.5% | 97.0% | 97.5% | 99.5% | 99.5% | 98.6% |