| Literature DB >> 35808522 |
Paul Thiry1,2,3, Martin Houry4, Laurent Philippe4, Olivier Nocent5, Fabien Buisseret3,6, Frédéric Dierick3,7,8, Rim Slama9, William Bertucci5, André Thévenon2, Emilie Simoneau-Buessinger1.
Abstract
Nowadays, the better assessment of low back pain (LBP) is an important challenge, as it is the leading musculoskeletal condition worldwide in terms of years of disability. The objective of this study was to evaluate the relevance of various machine learning (ML) algorithms and Sample Entropy (SampEn), which assesses the complexity of motion variability in identifying the condition of low back pain. Twenty chronic low-back pain (CLBP) patients and 20 healthy non-LBP participants performed 1-min repetitive bending (flexion) and return (extension) trunk movements. Analysis was performed using the time series recorded by three inertial sensors attached to the participants. It was found that SampEn was significantly lower in CLBP patients, indicating a loss of movement complexity due to LBP. Gaussian Naive Bayes ML proved to be the best of the various tested algorithms, achieving 79% accuracy in identifying CLBP patients. Angular velocity of flexion movement was the most discriminative feature in the ML analysis. This study demonstrated that: supervised ML and a complexity assessment of trunk movement variability are useful in the identification of CLBP condition, and that simple kinematic indicators are sensitive to this condition. Therefore, ML could be progressively adopted by clinicians in the assessment of CLBP patients.Entities:
Keywords: artificial intelligence; inertial measurement unit—IMU; machine learning; movement complexity; sample entropy; trunk flexion
Mesh:
Year: 2022 PMID: 35808522 PMCID: PMC9269703 DOI: 10.3390/s22135027
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1NLBP subject in starting position for the b&r test. The targets on the stool and on the wall in front of the subject are shown (red points). A zoom on the sensors is shown in the inset of the figure (on the right). The respective coordinate system (X, Y, Z) used is shown for SENS1&2 and for SENS3.
Figure 2Typical traces of (A) Acc X (blue), Acc Y (purple), and Acc Z (orange) and (B) Gyr X (blue), Gyr Y (purple), and Gyr Z (orange) time series recorded with SENS1 during a b&r test in a healthy NLBP subject.
Classifier’s hyperparameters.
| ML Algorithm | Hyperparameters |
|---|---|
| BF KNN | number of neighbours ( |
| Linear SVM | regularization parameter ( |
| SVM RBF | C-parameter (0.001, 0.01, 0.1, |
| DT | maximum depth of the tree (1, 5, |
| RF | maximum depth of the tree (1, 5, |
| AdaBoost | maximum number of estimators at which boosting stops (5, 10, |
| GaussianNB | ratio of the largest variance of all features added to the variances for computational stability ( |
BF KNN: Brute-Force K-Nearest Neighbors, SVM: Support Vector Machine, RBF: radial basis function, DT: Decision Tree, RF: Random Forest, AdaBoost: Adaptive boosting, GaussianNB: Gaussian naive Bayes; hyperparameters in bold are the selected ones by the grid-search function.
Significative differences (T-test or Wilcoxon test depending on the normality or not of the distribution of the SampEn values) between CLBP and NLBP groups. All parameters are SampEn values defined in Section 2.3.
| SampEn | Gyr Y SENS1 | Gyr Z SENS2 | HCF | Gyr Y SENS2 | Acc X SENS2 | ||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| CLBP | NLBP | CLBP | NLBP | CLBP | NLBP | CLBP | NLBP | CLBP | NLBP | ||
| Mean | 0.161 | 0.208 | 0.625 | 0.516 | 0.272 | 0.326 | Median | 0.217 | 0.282 | 0.266 | 0.389 |
| SD | 0.05 | 0.072 | 0.168 | 0.144 | 0.053 | 0.100 | Q1 | 0.187 | 0.220 | 0.227 | 0.312 |
| SEM | 0.011 | 0.016 | 0.038 | 0.032 | 0.012 | 0.022 | Q3 | 0.261 | 0.343 | 0.407 | 0.523 |
| 0.021 | 0.035 | 0.044 | 0.021 | 0.047 | |||||||
| Difference NLBP−CLBP | |||||||||||
| Mean | 0.035 | −0.108 | 0.055 | Mean | −0.034 | 0.097 | |||||
| SD | 0.168 | 0.254 | 0.111 | SD | 0.159 | 0.301 | |||||
| CI | 0.074 | 0.111 | 0.046 | CI | 0.070 | 0.132 | |||||
| SEM | 0.038 | 0.057 | 0.024 | SEM | 0.036 | 0.067 | |||||
| MDC | 0.104 | 0.157 | 0.070 | MDC | 0.099 | 0.187 | |||||
CI: 95% confidence interval; SEM: Standard Error of Measure; MDC: Minimal Detectable Change; SampEn value for: the Y-axis of the gyroscope from the sensor 1 (Gyr Y SENS1), the Z-axis of the gyroscope from the sensor 2 (Gyr Z SENS2), Hip Complexity Factor (HCF), the Y-axis of the gyroscope from the sensor 2 (Gyr Y SENS2), the X-axis of the accelerometer from the sensor 2 (Acc X SENS2).
Figure 3Principle of cycle segmentation based on minima of Acc Z. The dashed horizontal line represents the global threshold (40% above the global minima). The dotted vertical lines represent the cycle limits that lie in the middle of two consecutive local minima. The pink lines show the part of the time series that is below the threshold, the orange lines show the time series that is above the threshold, and the blue dots show the minima of the pink lines.
Figure 4Mean Gyr Y of the SENS2 cycle as function of time is shown for NLBP subjects (green dashed dot line) and for CLBP patients (red dashed dot line). The colored areas correspond to the cycles SD for healthy NLBP subjects (green) and CLBP patients (red or brown when overlapped with green). Note that all cycles were normalized to the same number of points (n = 450) and that mean and SD refer to all cycle values at a given normalized time point (itime).
Comparison of prediction performance between the whole sequences and cycle segmentation procedures, for all considered ML algorithms.
| Whole Sequences | Cycle Segmentation | |||
|---|---|---|---|---|
| Algorithms | Accuracy (%) | AUC | Accuracy (%) | AUC |
| BF KNN | 0.63 ± 0.08 | 0.69 ± 0.09 | 0.65 ± 0.05 | 0.67 ± 0.06 |
| Linear SVM | 0.72 ± 0.07 | 0.79 ± 0.07 | 0.68 ± 0.06 | 0.71 ± 0.08 |
| SVM RBF | 0.52 ± 0.06 | 0.52 ± 0.09 | 0.64 ± 0.04 | 0.71 ± 0.06 |
| DT | 0.66 ± 0.08 | 0.65 ± 0.09 | 0.66 ± 0.06 | 0.65 ± 0.06 |
| RF | 0.78 ± 0.07 | 0.83 ± 0.08 | 0.72 ± 0.05 | 0.80 ± 0.06 |
| AdaBoost | 0.68 ± 0.07 | 0.74 ± 0.08 | 0.70 ± 0.06 | 0.74 ± 0.08 |
| GaussianNB |
|
| 0.69 ± 0.07 | 0.74 ± 0.07 |
BF KNN: Brute-Force K-Nearest Neighbors, SVM: Support Vector Machine, RBF: radial basis function, DT: Decision Tree, RF: Random Forest, AdaBoost: Adaptive boosting, GaussianNB: Gaussian Naive Bayes. Bold numbers indicate best prediction results.
Accuracy and AUC scores for CLBP-NLBP classification using the whole sequences with different features.
| Whole Sequences | Raw Data | SampEn | CF | |||
|---|---|---|---|---|---|---|
| Algorithms | Accuracy (%) | AUC | Accuracy (%) | AUC | Accuracy (%) | AUC |
| BF KNN | 0.63 ± 0.08 | 0.69 ± 0.09 | 0.59 ± 0.10 | 0.62 ± 0.09 | 0.73 ± 0.06 | 0.78 ± 0.06 |
| Linear SVM | 0.72 ± 0.07 | 0.79 ± 0.07 | 0.53 ± 0.02 | 0.64 ± 0.10 | 0.68 ± 0.06 | 0.74 ± 0.07 |
| SVM RBF | 0.52 ± 0.06 | 0.52 ± 0.09 | 0.53 ± 0.02 | 0.64 ± 0.10 |
|
|
| DT | 0.66 ± 0.08 | 0.65 ± 0.09 | 0.58 ± 0.07 | 0.56 ± 0.07 | 0.60 ± 0.10 | 0.61 ± 0.10 |
| RF | 0.78 ± 0.07 | 0.83 ± 0.08 | 0.59 ± 0.08 | 0.64 ± 0.09 | 0.68 ± 0.07 | 0.71 ± 0.07 |
| AdaBoost | 0.68 ± 0.07 | 0.74 ± 0.08 | 0.55 ± 0.10 | 0.57 ± 0.10 | 0.62 ± 0.10 | 0.62 ± 0.11 |
| GaussianNB |
|
|
|
| 0.60 ± 0.08 | 0.60 ± 0.10 |
BF KNN: Brute-Force K-Nearest Neighbors, SVM: Support Vector Machine, RBF: radial basis function, DT: Decision Tree, RF: Random Forest, AdaBoost: Adaptive boosting, GaussianNB: Gaussian Naive Bayes. Bold numbers indicate best prediction results.
Number of times the most discriminating characteristics are first and second out of 700 runs.
| Feature | First | Feature | Second |
|---|---|---|---|
| Gyr Y SENS2 min | 355 | Acc Y SENS2 SD | 114 |
| Acc X SENS3 Q3 | 136 | Gyr Y SENS2 min | 112 |
| Acc Y SENS2 SD | 111 | Acc X SENS3 Q3 | 103 |
| Acc Y SENS2 SD | 89 | Acc Y SENS2 Q1 | 83 |
| Acc X SENS3 Q1 | 64 | Acc X SENS3 mean | 71 |
Gyr: Gyroscope; Acc: Accelerometer; SENS: Sensor; min: minimum; SD: standard deviation; Q1: 1st quartile; Q3 3rd quartile.