| Literature DB >> 34178951 |
Panyawut Sri-Iesaranusorn1, Attawit Chaiyaroj2, Chatchai Buekban2, Songphon Dumnin2, Ronachai Pongthornseri2, Chusak Thanawattano2, Decho Surangsrirat2.
Abstract
Surface electromyography (sEMG) is a non-invasive and straightforward way to allow the user to actively control the prosthesis. However, results reported by previous studies on using sEMG for hand and wrist movement classification vary by a large margin, due to several factors including but not limited to the number of classes and the acquisition protocol. The objective of this paper is to investigate the deep neural network approach on the classification of 41 hand and wrist movements based on the sEMG signal. The proposed models were trained and evaluated using the publicly available database from the Ninapro project, one of the largest public sEMG databases for advanced hand myoelectric prosthetics. Two datasets, DB5 with a low-cost 16 channels and 200 Hz sampling rate setup and DB7 with 12 channels and 2 kHz sampling rate setup, were used for this study. Our approach achieved an overall accuracy of 93.87 ± 1.49 and 91.69 ± 4.68% with a balanced accuracy of 84.00 ± 3.40 and 84.66 ± 4.78% for DB5 and DB7, respectively. We also observed a performance gain when considering only a subset of the movements, namely the six main hand movements based on six prehensile patterns from the Southampton Hand Assessment Procedure (SHAP), a clinically validated hand functional assessment protocol. Classification on only the SHAP movements in DB5 attained an overall accuracy of 98.82 ± 0.58% with a balanced accuracy of 94.48 ± 2.55%. With the same set of movements, our model also achieved an overall accuracy of 99.00% with a balanced accuracy of 91.27% on data from one of the amputee participants in DB7. These results suggest that with more data on the amputee subjects, our proposal could be a promising approach for controlling versatile prosthetic hands with a wide range of predefined hand and wrist movements.Entities:
Keywords: Ninapro database; deep neural network; hand movement classification; prosthetic hand; surface electromyogram
Year: 2021 PMID: 34178951 PMCID: PMC8220079 DOI: 10.3389/fbioe.2021.548357
Source DB: PubMed Journal: Front Bioeng Biotechnol ISSN: 2296-4185
Figure 1Overview of hand movement classification using sEMG.
Overview of the datasets in the Ninapro project.
| Intact participants | 27 | 40 | – | 10 | 10 | 10 | 20 | 10 |
| Amputees | – | – | 11 | – | – | – | 2 | 2 |
| Repetitions | 10 | 6 | 6 | 6 | 6 | 70 | 6 | 12 |
| Movements | 53 | 50 | 50 | 53 | 53 | 7 | 41 | 9 |
| Sensors | Otto | Delsys | Delsys | Cometa | Myo | Delsys | Delsys | Delsys |
| Sampling rate | 100 Hz | 2 kHz | 2 kHz | 2 kHz | 200 Hz | 2 kHz | 2 kHz | 2 kHz |
Figure 2Illustration of the 41 movements in this study according to the grouping from the Ninapro project: isometric and isotonic hand configurations and basic wrist movements (17 exercises), grasping and functional movements (23 exercises), and rest position.
Figure 3Distributions of the sEMG signal amplitudes grouped by different experiment conditions. The first two rows show data from the intact and amputee groups from DB7, while the third row shows DB5. The first column groups the amplitudes from all movements and subjects by repetition; the second column from all repetitions and subjects by movement; the third column from all repetitions and movements by subject. The horizontal lines in the middle of each box mark the median; the edges denote the first and third quartiles; the whiskers cover approximately 2.7 times the standard deviation.
Figure 4Balanced classification accuracy at different thresholds for the intact and amputee groups from DB7.
Figure 5Schematic of the proposed deep neural network model. The sEMG input is segmented by a sliding window. Then, the features are extracted and normalized before passing into the classifier. Lastly, a softmax activation function turns the classifier's output into a probability-like vector for the classification of 41 movements.
Figure 6(A) Overall accuracy and (B) balanced accuracy of the proposed deep neural network classifiers for intact participants; (C) overall accuracy and (D) balanced accuracy of the classifiers for amputee participants. Note that (C,D) do not show standard deviation as there is only one subject shown in each graph.
Hand movement classes from DB5 and DB7 intact participants, and from DB7 amputees #1 and #2, grouped by recall.
| ≥ 0.8 | 30 classes: 0, 1, 2, 3, 4, | 31 classes: 0, 1, 2, 3, 4, |
| 5, 6, 7, 8, 9, 10, 13, 14, 15, | 5, 6, 7, 8, 9, 10, 11, 12, 13, | |
| 16, 17, 21, 23, 25, 27, 28, 29, | 14, 15, 16, 17, 19, 21, 23, 25, | |
| 32, 34, 35, 36, 37, 38, 39, 40 | 27, 32, 34, 35, 36, 37, 38, 39, 40 | |
| <0.6 | 11 classes: 11, 12, 18, 19, | 10 classes:18, 20, 22, 24, |
| 20, 22, 24, 26, 30, 31, 33 | 26, 28, 29, 30, 31, 33 | |
| ≥ 0.8 | 10 classes: 0, 3, 4, 5, | 22 classes: 0, 2, 5, 6, 9, 10, |
| 6, 10, 16, 27, 35, 38 | 11, 13, 14, 15,16, 17, 18, 22, | |
| 24, 25, 32, 34, 36, 37, 39, 40 | ||
| 0.6 – 0.8 | 15 classes: 1, 9, 11, | 10 classes: 1, 7, 19, 20, |
| 12, 13, 14, 17, 18, 19, | 21, 27, 28, 31, 35, 38 | |
| 20, 22, 25, 28, 33, 34 | ||
| <0.6 | 14 classes: 2, 7, 8, | 9 classes: 3, 4, 8, 12 |
| 15, 21, 23, 24, 26, | 23, 26, 29, 30, 33 | |
| 29, 30, 31, 32, 36, 37 | ||
Figure 7Confusion matrices for the proposed model trained on data from amputees #1 (A) and #2 (B).
Figure 8Performance comparison with baseline studies for the classification of 41 hand movements. (A) Overall accuracy of our model on DB5 compared to Pizzolato et al. (2017). (B) Balanced accuracy on DB7 compared to Krasoulis et al. (2017). The standard deviations for the baseline prior works using eight channels of DB5 and 256 ms of DB7 were not provided.
Figure 9Overall accuracy (A) and Balanced accuracy (B) comparison between CNN and DNN with hand-crafted features for DB5 with 16 channels, intact group of DB7, and amputee group of DB7.
Figure 10Predicted sequences of full repetitions for every intact participant from DB7 (A) the large-diameter grasp (class 18) (B) the prismatic pinch grab (class 31).
Accuracy and macro-averaged metrics of the proposed deep neural network classifiers for 41 hand movements.
| 100 ms | 74.00 ± 2.10 | 42.79 ± 4.40 | 36.14 ± 4.80 | 38.68 ± 4.70 | |
| 200 ms | 77.97 ± 2.09 | 51.58 ± 4.73 | 46.13 ± 4.56 | 47.95 ± 4.69 | |
| 400 ms | 80.88 ± 1.99 | 56.44 ± 4.12 | 52.48 ± 4.55 | 53.85 ± 4.49 | |
| 800 ms | 87.04 ± 1.83 | 68.88 ± 4.08 | 67.56 ± 4.01 | 68.00 ± 4.06 | |
| 1,000 ms | 89.00 ± 2.05 | 73.32 ± 4.11 | 71.78 ± 4.67 | 72.35 ± 4.54 | |
| 100 ms | 81.37 ± 2.17 | 59.43 ± 5.13 | 55.10 ± 5.00 | 56.90 ± 5.03 | |
| 200 ms | 84.25 ± 2.02 | 65.21 ± 4.48 | 62.07 ± 4.56 | 63.38 ± 4.57 | |
| 400 ms | 87.21 ± 1.86 | 71.69 ± 3.68 | 68.54 ± 4.43 | 69.70 ± 4.31 | |
| 800 ms | 90.72 ± 1.62 | 77.45 ± 3.34 | 76.88 ± 3.84 | 76.88 ± 3.75 | |
| 1,000 ms | 93.87 ± 1.49 | 85.57 ± 2.46 | 84.00 ± 3.40 | 84.67 ± 3.20 | |
| 100 ms | 82.83 ± 4.90 | 73.96 ± 3.21 | 66.73 ± 4.89 | 69.78 ± 4.05 | |
| 200 ms | 85.08 ± 4.83 | 78.83 ± 2.88 | 70.67 ± 5.47 | 74.18 ± 4.38 | |
| 400 ms | 87.74 ± 4.94 | 81.81 ± 3.29 | 76.56 ± 5.31 | 78.90 ± 4.47 | |
| 800 ms | 90.61 ± 4.73 | 85.48 ± 3.55 | 82.30 ± 5.10 | 83.77 ± 4.50 | |
| 1,000 ms | 91.69 ± 4.68 | 87.03 ± 4.06 | 84.66 ± 4.78 | 85.74 ± 4.55 | |
| 100 ms | 74.64 | 55.27 | 48.20 | 50.18 | |
| 200 ms | 78.23 | 63.91 | 53.79 | 56.05 | |
| 400 ms | 79.16 | 64.55 | 57.56 | 58.44 | |
| 800 ms | 81.54 | 69.97 | 62.02 | 63.08 | |
| 1000 ms | 82.42 | 72.84 | 65.10 | 65.66 | |
| 100 ms | 87.49 | 63.62 | 59.23 | 60.26 | |
| 200 ms | 89.68 | 69.20 | 65.33 | 66.23 | |
| 400 ms | 89.01 | 64.25 | 62.39 | 61.66 | |
| 800 ms | 93.41 | 76.39 | 74.87 | 74.35 | |
| 1,000 ms | 94.07 | 75.78 | 76.55 | 74.90 | |
Accuracy and macro-averaged metrics of the proposed deep neural network classifiers for SHAP prehensile patterns.
| 100 ms | 92.49 ± 1.26 | 69.16 ± 4.13 | 61.21 ± 6.18 | 64.71 ± 5.00 | |
| 200 ms | 93.71 ± 1.11 | 72.81 ± 4.26 | 68.59 ± 5.72 | 70.52 ± 4.79 | |
| 400 ms | 95.53 ± 1.24 | 80.56 ± 5.58 | 77.42 ± 6.24 | 78.76 ± 5.91 | |
| 800 ms | 97.29 ± 0.86 | 87.17 ± 2.90 | 86.27 ± 4.69 | 86.61 ± 3.94 | |
| 1,000 ms | 97.78 ± 0.75 | 89.14 ± 3.58 | 88.61 ± 3.95 | 88.93 ± 3.48 | |
| 100 ms | 94.66 ± 1.05 | 79.40 ± 3.23 | 74.14 ± 5.50 | 76.63 ± 4.22 | |
| 200 ms | 95.86 ± 0.89 | 82.74 ± 2.78 | 80.96 ± 4.37 | 81.79 ± 3.55 | |
| 400 ms | 97.28 ± 0.81 | 89.49 ± 2.34 | 87.10 ± 4.01 | 88.19 ± 3.05 | |
| 800 ms | 98.38 ± 0.84 | 93.17 ± 3.59 | 92.10 ± 3.95 | 92.60 ± 3.80 | |
| 1,000 ms | 98.82 ± 0.58 | 94.93 ± 1.91 | 94.48 ± 2.55 | 94.67 ± 2.18 | |
| 100 ms | 92.69 ± 4.19 | 86.42 ± 3.96 | 74.08 ± 9.32 | 79.40 ± 7.32 | |
| 200 ms | 93.34 ± 4.41 | 88.01 ± 4.02 | 76.81 ± 10.05 | 81.66 ± 7.86 | |
| 400 ms | 94.22 ± 4.03 | 88.90 ± 4.00 | 79.92 ± 8.91 | 83.77 ± 7.28 | |
| 800 ms | 95.46 ± 4.16 | 90.30 ± 4.38 | 85.18 ± 8.90 | 87.59 ± 7.03 | |
| 1,000 ms | 95.78 ± 3.99 | 90.94 ± 3.90 | 86.00 ± 8.35 | 88.31 ± 6.79 | |
| 100 ms | 88.29 | 68.02 | 56.73 | 59.82 | |
| 200 ms | 88.88 | 69.94 | 56.66 | 58.09 | |
| 400 ms | 90.20 | 76.42 | 63.87 | 65.71 | |
| 800 ms | 91.14 | 77.48 | 67.39 | 69.22 | |
| 1,000 ms | 92.53 | 83.22 | 71.71 | 74.97 | |
| 100 ms | 97.13 | 87.43 | 80.08 | 81.83 | |
| 200 ms | 97.55 | 87.11 | 83.99 | 83.38 | |
| 400 ms | 98.04 | 91.95 | 84.44 | 85.47 | |
| 800 ms | 98.66 | 94.93 | 88.92 | 89.92 | |
| 1,000 ms | 99.00 | 96.24 | 91.27 | 93.11 | |
Average prediction time per sample of the proposed deep neural network model.
| 100 ms | 35.89 |
| 200 ms | 53.43 |
| 400 ms | 76.75 |
| 800 ms | 89.85 |
| 1,000 ms | 98.86 |