| Literature DB >> 32977650 |
Xinrui Jiang1, Ye Zhang1, Qi Yang1, Bin Deng1, Hongqiang Wang1.
Abstract
At present, there are two obvious problems in radar-based gait recognition. First, the traditional radar frequency band is difficult to meet the requirements of fine identification with due to its low carrier frequency and limited micro-Doppler resolution. Another significant problem is that radar signal processing is relatively complex, and the existing signal processing algorithms are poor in real-time usability, robustness and universality. This paper focuses on the two basic problems of human gait detection with radar and proposes a human gait classification and recognition method based on millimeter-wave array radar. Based on deep-learning technology, a multi-channel three-dimensional convolution neural network is proposed on the basis of improving the residual network, which completes the classification and recognition of human gait through the hierarchical extraction and fusion of multi-dimensional features. Taking the three-dimensional coordinates, motion speed and intensity of strong scattering points in the process of target motion as network inputs, multi-channel convolution is used to extract motion features, and the classification and recognition of typical daily actions are completed. The experimental results show that we have more than 92.5% recognition accuracy for common gait categories such as jogging and normal walking.Entities:
Keywords: feature fusion; human gait recognition; millimeter-wave array radar; multi-channel three-dimensional convolution neural network
Mesh:
Year: 2020 PMID: 32977650 PMCID: PMC7582452 DOI: 10.3390/s20195466
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Geometric diagram of the relationship between the direction of arrival and array structure.
Direction of arrival (DOA) estimation method.
|
|
|---|
| calculate |
| end for |
Figure 2Radar system: (a) Structure of radar system; (b) Antenna array element and virtual aperture.
Figure 3Point cloud collection.
Figure 4Diagram of convolution process [32]: (a) 2-D convolution; (b) 2-D convolution on multiple frames; (c) 3-D convolution.
Figure 5Structure of multi-channel three-dimensional convolution neural network (MC-3DCNN).
Figure 6Experimental scenario: (a) Schematic diagram of experimental scene; (b) Real experimental scene.
Samples and labels.
| Category of Action | Label of Action | Number of Samples (Frame) |
|---|---|---|
| normal walking | 0 | 40,000 |
| jogging | 1 | 40,000 |
| lame walking | 2 | 40,000 |
| squatting down and standing up | 3 | 40,000 |
Figure 7Flow of data processing.
Capacity of the dataset.
| Dataset | Number (Frame) |
|---|---|
| Dataset 1 | 10,000 × 4 |
| Dataset 2 | 10,000 × 4 |
| Dataset 3 | 10,000 × 4 |
| Dataset 4 | 10,000 × 4 |
Cross-training and dataset partition.
| Cross-Validation | Dataset 1 | Dataset 2 | Dataset 3 | Dataset 4 |
|---|---|---|---|---|
| CV_1 | Training | Training | Training | Testing |
| CV_2 | Training | Training | Testing | Training |
| CV_3 | Training | Testing | Training | Training |
| CV_4 | Testing | Training | Training | Training |
Figure 8The structure of single-channel three-dimensional convolutional neural network (SC-3DCNN).
Figure 9Training loss during cross-training.
Figure 10Training accuracy during cross-training.
Recognition results of cross validation.
| Accuracy (%) | CV_1 | CV_2 | CV_3 | CV_4 | Average Accuracy | |
|---|---|---|---|---|---|---|
| Category | ||||||
| Jogging | 95.20 | 89.80 | 94.60 | 90.40 | 92.50 | |
| Normal walking | 90.40 | 95.20 | 96.80 | 89.60 | 93.00 | |
| Lame walking | 81.60 | 94.20 | 89.60 | 85.60 | 87.75 | |
| Squatting down and standing up | 94.80 | 92.50 | 89.80 | 93.60 | 92.68 | |
Recognition results for different networks.
| Accuracy (%) | CV_1 | CV_2 | CV_3 | CV_4 | Average Accuracy | |
|---|---|---|---|---|---|---|
| Network | ||||||
| SC-3DCNN | 84.20 | 89.80 | 81.60 | 89.60 | 86.30 | |
| MC-3DCNN | 90.50 | 92.93 | 92.70 | 89.80 | 93.00 | |