| Literature DB >> 35808358 |
Jungi Kim1, Haneol Seo2, Muhammad Tahir Naseem2, Chan-Su Lee3.
Abstract
Walking is an exercise that uses muscles and joints of the human body and is essential for understanding body condition. Analyzing body movements through gait has been studied and applied in human identification, sports science, and medicine. This study investigated a spatiotemporal graph convolutional network model (ST-GCN), using attention techniques applied to pathological-gait classification from the collected skeletal information. The focus of this study was twofold. The first objective was extracting spatiotemporal features from skeletal information presented by joint connections and applying these features to graph convolutional neural networks. The second objective was developing an attention mechanism for spatiotemporal graph convolutional neural networks, to focus on important joints in the current gait. This model establishes a pathological-gait-classification system for diagnosing sarcopenia. Experiments on three datasets, namely NTU RGB+D, pathological gait of GIST, and multimodal-gait symmetry (MMGS), validate that the proposed model outperforms existing models in gait classification.Entities:
Keywords: gait classification; global average pooling (GAP); graph convolutional networks (GCN); multiple-input branches (MIB); spatiotemporal graph convolutional networks (ST-GCN); temporal convolutional network (TCN)
Mesh:
Year: 2022 PMID: 35808358 PMCID: PMC9269520 DOI: 10.3390/s22134863
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Comparison and weaknesses of related work.
| Publications | Method | Dataset | Accuracy | Weakness |
|---|---|---|---|---|
| Khokhlova et al. [ | Using single LSTM | MMGS dataset | 94 | Lack of diverse datasets |
| Jun et al. [ | Using GRU | Newly created dataset | 93.7 | Lack of diverse datasets |
| Yan et al. [ | Using spatiotemporal GCN | Kinematics + NTU-RGBD | 88.3 | Less accuracy |
| Liao et al. [ | Using CNN | CASIA B + CASIA E | - | Uses few handcrafted features |
| Shi et al. [ | Using GCN | NTU-RGB + Kinematics skeleton | 90 | Unable to fuse RGB modality |
| Lie et al. [ | Using pose-refinement GCN | Kinematics + NTURGB-D | 91.7 | Less accuracy |
| Ding et al. [ | Using Semantics guided GCN | NTU + Kinetics | 94.2 | Use a smaller number of classes |
| Song et al. [ | Using multi-stream GCN | NTU RGB-D 60 + | 82.7 | Network complexities |
| Shi et al. [ | Using two-stream adaptive GCN | NTU RGB D + Kinetics | 95.1 | Network complexities |
| Si et al. [ | Using attention-enhanced GC LSTM | NTU RGB D + North-Western UCLA | 93.3 | Unable to fuse skeleton sequence with object appearance |
Figure 1Sample images from NTU RGB+D dataset [50].
Figure 2Skeleton data of normal and pathological gaits.
Figure 3Overall pipeline of our proposed approach.
Figure 4Overview of proposed model. Two numbers in each block denote input and output channels, and ⊕ represents concatenation.
Figure 5Depth of proposed ST-GCN block: details of ST-GCN (left) and spatial features of GCN and TCN (right).
Figure 6Overview of the proposed ST-GCN module, where C, T, and V denote the numbers of input channels, frames, and joints, respectively.
Performance for multiple data inputs for joints, velocity, and bones.
| Input Type | Accuracy (%) |
|---|---|
| Joints | 82.9 |
| Velocity | 82.4 |
| Bones | 84.0 |
| Joints + Velocity | 86.8 |
| Joints + Bones | 85.4 |
| Velocity + Bones | 87.3 |
| Joints + Velocity + Bones |
|
Comparison of performance of the proposed method with the schemes in the literature for NTU RGB+D dataset.
| Model | Accuracy (%) |
|---|---|
| ST-GCN [ | 81.5 |
| PR-GCN [ | 85.2 |
| Sem-GCN [ | 86.2 |
| AS-GCN [ | 86.8 |
| RA-GCN [ | 87.3 |
| PB-GCN [ | 87.5 |
| 2s-AGCN [ | 88.5 |
| AGC-LSTM [ | 89.2 |
| Multiple-input ST-GCN |
|
| Multiple-input ST-GCN |
|
Comparison of performance of proposed method with the schemes in the literature for pathological-gait dataset.
| Model | Accuracy (%) |
|---|---|
| GRU (full-skeleton) [ | 90.1 |
| GRU (only legs) [ | 93.7 |
| ST-GCN | 94.5 |
| Multiple-input ST-GCN |
|
| Multiple-input ST-GCN |
|
Comparison of performance of the proposed method with the schemes in the literature for MMGS dataset.
| Model | Accuracy (%) |
|---|---|
| Single-LSTM model [ | 94 |
| Ensemble-LSTM model [ | 91 |
| AGS-GCN [ |
|
Figure 7Separable convolution for skeleton-based action recognition.