| Literature DB >> 35721406 |
Zhipeng Gao1, Junyi Wu1, Tingting Wu1, Renyu Huang1, Anguo Zhang2,3, Jianqiang Zhao1.
Abstract
Recently, gait has been gathering extensive interest for the non-fungible position in applications. Although various methods have been proposed for gait recognition, most of them can only attain an excellent recognition performance when the probe and gallery gaits are in a similar condition. Once external factors (e.g., clothing variations) influence people's gaits and changes happen in human appearances, a significant performance degradation occurs. Hence, in our article, a robust hybrid part-based spatio-temporal feature learning method is proposed for gait recognition to handle this cloth-changing problem. First, human bodies are segmented into the affected and non/less unaffected parts based on the anatomical studies. Then, a well-designed network is proposed in our method to formulate our required hybrid features from the non/less unaffected body parts. This network contains three sub-networks, aiming to generate features independently. Each sub-network emphasizes individual aspects of gait, hence an effective hybrid gait feature can be created through their concatenation. In addition, temporal information can be used as complement to enhance the recognition performance, a sub-network is specifically proposed to establish the temporal relationship between consecutive short-range frames. Also, since local features are more discriminative than global features in gait recognition, in this network a sub-network is specifically proposed to generate features of local refined differences. The effectiveness of our proposed method has been evaluated by experiments on the CASIA Gait Dataset B and OU-ISIR Treadmill Gait Dataset B. Related experiments illustrate that compared with other gait recognition methods, our proposed method can achieve a prominent result when handling this cloth-changing gait recognition problem. ©2022 Gao et al.Entities:
Keywords: Clothing-independent; Gait recognition; Part-based; Spatio-temporal feature learning
Year: 2022 PMID: 35721406 PMCID: PMC9202625 DOI: 10.7717/peerj-cs.996
Source DB: PubMed Journal: PeerJ Comput Sci ISSN: 2376-5992
Figure 1Framework of the proposed method.
Figure 2Segmentation of human bodies.
Figure 3Sub-network used to extract local spatial features.
Figure 4Sub-network used to extract temporal features.
Sub-network parameters.
| Sub-network | Convolution channels | HPM scales |
|---|---|---|
| Sub-network 1 |
| {5, 4} |
| Sub-network 2 |
| {2, 4} |
| Sub-network 3 |
| 4 |
Comparison on CASIA-B under the same viewing angle by accuracies (%).
| Probe set |
|
|
|
|
|---|---|---|---|---|
| 36° (NM) |
|
| 90.5 | 89.5 |
| 36° (CL) |
|
| 90.9 | 91.1 |
| 54° (NM) |
|
| 91.1 | 88.2 |
| 54° (CL) |
|
| 93.2 | 91.9 |
| 72° (NM) |
|
| 94.7 | 88.7 |
| 72° (CL) |
|
| 96.5 | 89.5 |
| 90° (NM) |
|
| 93.5 | 87.1 |
| 90° (CL) |
|
| 95.1 | 88.7 |
| 108° (NM) |
|
| 92.7 | – |
| 108° (CL) |
|
| 94.1 | – |
| 126° (NM) |
|
| 91.1 | – |
| 126° (CL) |
|
| 91.5 | – |
| 144° (NM) |
|
| 92.2 | – |
| 144° (CL) |
|
| 93.5 | – |
Notes.
The first and second highest scores are represented by bold and underline, respectively.
Comparison on CASIA-B under different walking conditions by accuracies (%).
| (Probe, Gallery) |
|
|
|
|
|
|---|---|---|---|---|---|
| (36°, 54°) |
|
| 87.0 | 59.8 | 49.7 |
| (54°, 72°) |
|
| 90.0 | 72.5 | 62.0 |
| (72°, 90°) |
|
| 94.2 | 88.5 | 78.3 |
| (90°, 108°) |
|
| 86.5 | 85.7 | 75.6 |
| (108°, 126°) |
|
| 89.8 | 68.8 | 58.1 |
| (126°, 144°) |
|
| 91.2 | 62.5 | 51.4 |
| Mean |
|
| 89.8 | 73.0 | 62.5 |
Notes.
The first and second highest scores are represented by bold and underline, respectively.
Averaged rank-1 accuracies (%) on CASIA-B using setting LT, excluding identical-view cases.
| Gallery NM#1-4 | Probe views | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Probe CL#1-2 | 0° | 18° | 36° | 54° | 72° | 90° | 108° | 126° | 144° | 162° | 180° | Mean |
|
| 37.7 | 57.2 | 66.6 | 61.1 | 55.2 | 54.6 | 55.2 | 59.1 | 58.9 | 48.8 | 39.4 | 54.0 |
|
| 42.1 | – | – | 70.7 | – | 70.6 | – | 69.4 | – | – | – | 63.2 |
|
| 61.4 | 75.4 | 80.7 | 77.3 | 72.1 | 70.1 | 71.5 | 73.5 | 73.5 | 68.4 | 50.0 | 70.4 |
|
| 64.7 | 79.4 | 84.1 | 80.4 | 73.7 | 72.3 | 75.0 | 78.5 | 77.9 | 71.2 | 57.0 | 74.0 |
|
| 70.6 | 82.4 | 85.2 | 82.7 |
|
| 76.2 | 78.9 | 77.9 |
| 64.3 | 77.5 |
|
|
| 81.0 | 82.1 | 82.8 |
|
| 75.5 | 77.4 | 72.3 | 73.5 |
| 77.6 |
|
|
|
|
|
| 77.1 | 72.5 |
|
|
|
|
|
|
|
| 63.4 | 77.3 | 80.1 | 79.4 | 72.4 | 69.8 | 71.2 | 73.8 | 75.5 | 71.7 | 62.0 | 72.4 |
|
| 65.8 | 80.7 | 82.5 | 81.1 | 72.7 | 71.5 | 74.3 | 74.6 | 78.7 | 75.8 | 64.4 | 74.7 |
|
| 64.2 | 80.9 | 83.0 | 79.5 | 74.3 | 69.1 | 74.8 | 78.5 | 81.0 | 77.0 | 60.3 | 74.8 |
|
| 68.3 |
|
|
| 77.8 | 76.1 |
|
|
| 78.0 | 59.3 |
|
Notes.
The first and second highest scores are represented by bold and underline, respectively.
Figure 5Thirty-two clothing combinations used in OU-ISIR Treadmill Dataset B.
Comparison on OU-ISIR Treadmill Dataset B by accuracies (%).
| Probe set type |
|
|
|
|
|---|---|---|---|---|
| 0 |
|
| 94.0 |
|
| 2 |
|
| 93.5 |
|
| 3 |
|
| 91.6 |
|
| 4 |
|
| 94.1 | 98.5 |
| 5 |
|
| 94.5 | 94.1 |
| 6 |
|
| 92.0 | 91.2 |
| 7 |
|
| 94.2 | 94.1 |
| 8 |
|
| 94.5 | 94.1 |
| 9 |
|
| 92.0 | 97.1 |
| A |
|
| 91.6 | 91.2 |
| B |
|
| 88.2 | 95.6 |
| C |
|
| 94.5 | 94.1 |
| D |
|
| 92.0 |
|
| E |
|
| 91.5 | 91.2 |
| F |
|
| 93.1 |
|
| G |
|
| 89.1 | 98.5 |
| H |
|
| 95.0 | 94.1 |
| I |
|
| 98.5 | 98.5 |
| J |
|
| 91.5 | 91.2 |
| K |
|
| 87.5 | 98.5 |
| L |
|
| 90.0 |
|
| M |
|
| 97.5 | 97.1 |
| N |
|
| 85.5 |
|
| P |
|
| 91.1 |
|
| R |
|
| 86.2 | 88.2 |
| S |
|
| 89.1 | 95.6 |
| T |
|
| 95.0 | 94.1 |
| U |
|
| 95.5 | 94.1 |
| V |
|
| 91.6 | 91.2 |
| X |
|
| 90.1 |
|
| Y |
|
| 89.0 |
|
| Z |
|
| 87.2 | 98.5 |
Notes.
The first and second highest scores are represented by bold and underline, respectively.
Effectiveness of different input frames.
| Input frame number | Accuracy (%) |
|---|---|
| (5, 5) | 23.1 |
| (10, 10) | 73.8 |
| (15, 15) | 77.0 |
| (20, 20) | 77.3 |
| (25, 25) | 78.2 |
| (30, 30) | 78.5 |
| (35, 35) | 77.9 |
Effectiveness of different sub-networks.
| Feature component | Accuracy (%) |
|---|---|
|
| 68.4 |
|
| 77.9 |
|
| 69.6 |
|
|