| Literature DB >> 35047564 |
Elsa J Harris1, I-Hung Khoo2,3, Emel Demircan1,3.
Abstract
We performed an electronic database search of published works from 2012 to mid-2021 that focus on human gait studies and apply machine learning techniques. We identified six key applications of machine learning using gait data: 1) Gait analysis where analyzing techniques and certain biomechanical analysis factors are improved by utilizing artificial intelligence algorithms, 2) Health and Wellness, with applications in gait monitoring for abnormal gait detection, recognition of human activities, fall detection and sports performance, 3) Human Pose Tracking using one-person or multi-person tracking and localization systems such as OpenPose, Simultaneous Localization and Mapping (SLAM), etc., 4) Gait-based biometrics with applications in person identification, authentication, and re-identification as well as gender and age recognition 5) "Smart gait" applications ranging from smart socks, shoes, and other wearables to smart homes and smart retail stores that incorporate continuous monitoring and control systems and 6) Animation that reconstructs human motion utilizing gait data, simulation and machine learning techniques. Our goal is to provide a single broad-based survey of the applications of machine learning technology in gait analysis and identify future areas of potential study and growth. We discuss the machine learning techniques that have been used with a focus on the tasks they perform, the problems they attempt to solve, and the trade-offs they navigate.Entities:
Keywords: artificial intelligence; biometrics; human gait analysis; machine learning; review
Year: 2022 PMID: 35047564 PMCID: PMC8762057 DOI: 10.3389/frobt.2021.749274
Source DB: PubMed Journal: Front Robot AI ISSN: 2296-9144
FIGURE 1Number of papers by year of publication.
Smart gait vocabulary.
| SG Vocabulary | Definition/Context in our papers |
|---|---|
| Artificial Intelligence | Artificial Intelligence is a technology that enables computers and devices to act intelligently and make decisions like humans ( |
| Machine Learning (ML) | Machine Learning is a subfield of AI that enables computers and devices to learn from data without being explicitly programmed ( |
| Abnormal Gait Detection | The task of distinguishing a healthy gait from a pathological gait. Some of the pathologies that affect the walking pattern as discussed in this paper include dementia, Huntington’s disease (HD), PD, Autism Spectrum Disorder (ASD), Amyotrophic Lateral Sclerosis (ALS), Post-Stroke Hemiparetic (PSH), Acquired Brain Injury (ABI), depression, neuromuscular disease, lower extremity muscle fatigue, spastic diplegia, Cerebral Palsy (CP), etc. |
| Human Identification | Presented gait data is compared to a set of gait data with known identities (labeled training data) to determine whom the unknown gait belongs to |
| Human Re-identification | The task of identifying images of the same person from non-overlapping camera views at different times and locations. Gait is a behavioral biometric feature that is unobtrusive, hard to fake or conceal, and can be perceived from a distance without requiring the subject’s active collaboration ( |
| Fall Detection | A binary classification task, usually concurrent with activity recognition that classifies an activity as fall or no fall |
| Activity Recognition | A classification task that maps features extracted from various sensor raw data to classes corresponding to activities such as sitting, lying, running, walking, stair climbing |
| Gender Recognition | Gender Recognition is a binary classification that maps features to qualitative outputs: male and female |
| Smart Home | A smart home utilizes context-aware and location-aware technologies to create intelligent automation and ubiquitous computing home environment for comfort, energy management, safety, and security ( |
| Gait Event Detection | Detection of a sequence of events that specifies the transition from one gait phase to another during each gait cycle. ( |
| Kinetic and Kinematic analysis | Kinematics studies the motion of body segments without considering masses or causal forces. Kinetics studies the relation between motion and its causes |
| Biometric Authentication | An automated method of verifying a person’s identity based on their biometric (gait) characteristics |
| Crowd Density | The density level of people in a crowded scene |
| Anomaly detection | It labels a behavior pattern that is "far away" from a trained model as anomalous, where “far away” is measured by a time-varying threshold ( |
| Gait estimation from Pose | Parameters such as step length, stride length, stride time, cadence, etc., are estimated from the human pose |
| Human Gait Motion Modelling | A probabilistic manifold-based motion modeling framework able to model with a variety of walking styles from different individuals and with different strides ( |
| Occupant Activity Sensing | Actively knowing the identity of the people within a monitored area and what they are doing ( |
| Multi-Gait Recognition | Multi-gait is a term used by authors ( |
| Brain-Computer Interface (BCI) | A technology that translates signals from human brain activity such as walking intention to a command sent to an external assistive, adaptive, or rehabilitative device, such as a prosthetic leg ( |
| Hybrid BCI (hBCI) | A system that fuses two bio-signals, where at least one is intentionally controlled. The different signals, such as data from electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS), are processed in real-time to establish and maintain communication between the brain and the computer. The output is evaluated through a feedback control loop. Compared with systems that use one modality alone, hBCI improves classification accuracy and the number of control commands by integrating the complementary properties of different modalities and removing artifacts ( |
| Kinetic Energy Harvesting (KEH) | Technology that converts kinetic motion into energy. The individuality of the gait pattern can be captured in the output voltage signals of KEH systems, with the added benefit of energy savings, compared to accelerometers. Thus KEH systems are used as sensors and energy sources simultaneously ( |
| The Digital Human | Digital replicate of a human in the virtual space. Automatic, continuous gait monitoring will be an integral part of such systems ( |
| IoT | IoT is a ubiquitous system of objects that are connected to the network, uniquely identifiable, capable of collecting, communicating, and processing data and AI-enabled to make autonomous decisions, individually or collectively ( |
Gait phase recognition and gait event prediction.
| Reference | Algorithm | Input data | AI task |
|---|---|---|---|
|
| LSTM-Net, DENSE-Net | 12 subjects, 7 IMUs, 2 IMU pressure insoles | offline AL to reduce labeling cost, online gait phase classification |
|
| Hybrid SA/GA | 3 subjects, 1 HC, 2 impaired gaits, IMU at the back of the heel | online gait event detection |
|
| MLP | 23 subjects, 1 electro-goniometer per leg | gait phase classification, 2 phases: stance/swing |
|
| MLP | 23 healthy subjects, sEMG and barographic data, natural walking conditions | gait phase classification, 2 phases: stance/swing. Gait event prediction, HS/TO |
|
| LDA | 9 healthy subjects, TW, 8 pressure sensors in an ankle-worn band | wearable gait phase recognition system |
|
| DT, RF, MLP and SVM | 31 subjects, an inertial sensor at the thigh | gait event detection |
|
| novel DC w/hierarchical, weighted HMMs | 10 healthy subjects, TW, 2–3 IMU gyroscopes | gait phase recognition |
|
| HMM with STV | 9 healthy subjects, TW, 1 IMU gyro at the instep of left foot | online gait event detection |
Legend: Distributed Classifier (DC), Hidden Markov Models (HMM), Short-Time Viterbi (STV), Treadmill Walking (TW), Heel-Strike/Toe-Off (HS/TO), Simulated Annealing (SA), Genetic Algorithm (GA), Surface Electromyography (sEMG).
Summary of fatigued gait studies.
| Reference | Algorithm/best accuracy reported | How is data collected | Task |
|---|---|---|---|
|
| CNN 97.8% | accelerometer worn around the chest, GPS watch for location tracking, 1 person | HAR: Climb Gate/Lay/Sit/Walk/Run. Variations in terrain and fatigue |
|
| MHTSCA with DTW as a dissimilarity measure | IMU worn at the right ankle 15 subjects | fatigue development over time |
|
| RF with BSS 85.5% | one sensor in the torso, 15 subjects | 4-phase fatigue management framework in the workplace (1) detection (2) Identification (3) diagnosis: whole-body vs. localized (4) recovery |
|
| CNN/RNN | smartphone sensors, images and videos from the camera | detection of fatigue due to Covid-19 |
|
| 2-class SVM 91% | 24 subjects, smartphone attached to the shank | detection of fatigue: baseline, low, medium, and strong fatigued states |
| 3-class SVM 76% | |||
| 4-class SVM 61% | |||
|
| SVM 90% | one IMU in the ankle, 20 subjects | detection of fatigue after MMH tasks |
|
| SVM 96% | 17 subjects, IMU at sternum level | recognition of localized fatigued/non-fatigued state |
|
| SVM and SOM with PCA 98.1% | 9 subjects GRFs | inter and intra-personal gait classification before, during, and after leg exhaustion |
Legend: Best Subset Selection (BSS), Manual Material Handling (MMH), Multivariate Hierarchical Time Series Clustering Algorithm (MHTSCA).
The pathological gait.
| Reference | Algorithm/Best Accuracy | Data Collection/Input | Pathology/Task Output |
|---|---|---|---|
|
| SVM with PCA 88.89% | Kinect camera with image rectification | Automatic depression detection |
|
| HealthXAI CART | Partial CASAS dataset, 192 subjects: 19 PwD, 54 MCI | Numerical score and explanation of the decline of cognitive functions of the elderly |
|
| ANN 93.9% | IMU at the waist belt, | Stroke prognostic tool, able/unable to return to work |
|
| Bidirectional GRU | GaitNDD, GaitPDB. Streaming of live and historical GRFs | ND: PD, HD, ALS gait normality analysis |
|
| Novel DNN with t-SNE F-score: 97.33% | Own dataset: UOM-GAIT-69. Tomography floor sensor raw data. | Age-related differences in healthy adults undertaking dual tasks |
|
| RF with IAFSA RMSE = 0.073 | 3 patients with knee replacement. Public dataset/challenge2 | Knee joint impairment KFC prediction |
|
| Kernel PCA with SVM, RF, ANN: 90% |
| Geriatric condition |
|
| Deep CNN AUC = 0.87 | DREAM PDDB Challenge | PD vs healthy gait; Large scale screening |
|
| RBF neural network with DL 95.61% | Kinematic modeling using a biped, | Chronic unilateral ACL deficiency. Classify ACL-D/ACL-I knees |
|
| Novel FL Sp = 95.2%, Se = 84.9% | Smartphone data | Real time, interpretable FoG detection |
|
| MLR |
| Effects of inflammation on human gait |
|
| LR, SVM, RF | 4 people with MS. GRFs from instrumented treadmill | GML4MS framework, HC/MS mild and moderate classifier |
|
| ST-GCN and CVAE 88% | 4,277 human gaits in video and synthetic gaits by novel STEP-Gen | Emotion classification: happy, sad, angry, or neutral |
|
| WeedGait, by LSTM and SVM 92.1% |
| assesses marijuana-induced gait impairment passively, warns against DUIM online |
|
| SVM and BiLSTM |
| normal, in-toeing, out-toeing, and drop-foot gait |
| (Zhang et al., 2019b) | ANN (a = 50) 93.5% |
| Gait classification for CP patients with spastic diplegia |
|
| ST-ACF DTW, KNN with OpenPose | CASIA-B dataset. Frontal videos of two PD patients | Quantifying normal and Parkinsonian gait features from home movies |
|
| RF 91.58% | 95 graduate students. 52 score-depressed, 43 HC. Two MS Kinect cameras | Depression analysis |
|
| Logitboost & RF 94.5% on raw data |
| HD gait classification |
|
| ANFIS/PSO with LOOCV. 94.44% | 64 subjects, ALS = 13, PD = 15, HD = 20, HC = 16, ND Public dataset. Force-sensitive switches are placed on subjects’ shoes. | Classification of Gait Patterns in Patients with various ND |
|
| RF with DTW | 26 HSP and 33 healthy children. Optokinetic IGA system | Monitoring HSP progression and personalizing therapies |
|
| DMLP. 97.9% |
| Analyze speech and movement data captured by smartphone to estimate the severity of PD |
|
| ANN, SVM with SWDA 93.3% | 3D GRF data of 60 children: 30 ASD and 30 typically developing | Identifying ASD Gait |
|
| SVM w/PCA 98.21% |
| Recognition and Assessment of PSH Gait |
|
| DT. 93.75% | 49 YouTube videos of varying resolution. Video obtained through any pervasive devices | PD gait classification |
|
| LSTM HAR: 96.7% AAD: 91.43% | Public dataset collected in 3 households through environmental sensors ( | HAR and AAD for elderly people with dementia |
|
| LDA, NBC. 90.93% | GaitNDD. Force-sensitive resistors. 3 ALS, 15 PD, 20 HD, and 16 HC | Classification of ALS among other ND diseases and healthy subjects |
|
| GPLVM-thold and KNN-DTW F1-score > 0.94 |
| Discriminate between healthy and pathological gait patterns because of stroke or ABI |
|
| SVM with Gaussian RBF 83.3% | GaitNDD. GRF measurements. | Distinguish PD gait from HD, ALS, and HC |
|
| NBC 94.1% |
| PD diagnosis |
|
| RF with MR normalization. 92.6% |
| PD diagnosis and management using normalized spatial-temporal gait |
Legend: Decision Tree (DT), K-Nearest Neighbors (KNN), Center for Advanced Studies in Adaptive Systems (CASAS) (Cook et al., 2015), Classification and Regression Trees (CART), Gated Recurrent Unit (GRU), Root Mean Square Error (RMSE), Receiver-Operating Characteristic (ROC), ACL Deficient (ACL-D), ACL-intact (ACL-I), Radial Basis Function (RBF), Deterministic Learning (DL), Multivariable Linear Regression (MLR), Multiple Sclerosis (MS), Gait data-based ML framework for MS prediction (GML4MS), Linear Regression (LR), Abnormal Activity Detection (AAD), Gaussian Process (GP) Latent Variable Models (GPLVM), OpenPose (Cao et al., 2017).
Datasets: CASIA-B (Yu et al., 2006), Gait in Neurodegenerative Disease Database (GaitNDD) (Hausdorff et al., 2000), Gait in Parkinson’s Disease (GaitPDB) (Goldberger et al., 2000), CASAS (Cook et al., 2015), 2 https://simtk.org/projects/kneeloads, ND Public Dataset (Hausdorff et al., 2000).
SG in sports.
| Reference | Algorithm | Data Collection/Input | AI Task/Output |
|---|---|---|---|
|
| Linear SVM 96% |
| ACL risk prediction in female basketball players via LESS score |
|
| CNN, not enough accuracy | Wearable accelerometer | predict near real-time GRF/Ms from kinematic data |
|
| CNN | 7 IMU’s | Gait classification: athlete vs. foot abnormalities |
|
| TS-DBN | Public datasets of videos KTH and UCF | HAR/sports behavior recognition |
|
| CNN | shoe-mounted accelerometer | Abnormal running kinematics Activity recognition |
|
| DeepLabCut | single GoPro camera | Markerless 2D kinematic analysis of underwater running |
|
| FFT | Smartphone (unconstrained) | Detects walking, counts steps, irrespective of phone placement |
|
| ANN with IG | infrared cameras and force plates | Influence of shoe midsole resilience and upper structure on running kinematics and kinetics |
|
| KNN with DTW | pressure sensor mat | Exercise detection and exercise count |
Legend: Fast Fourier Transform (FFT), Time-Space Deep Belief Network (TS-DBN), Landing Error Score System (LESS), Ground Reaction Forces and Moments (GRF/M), DeepLabCut as in (Mathis et al., 2018).
Datasets: Royal Institute of Technology (KTH) (Jaouedi et al., 2020) and University of Central Florida (UCF) (Perera et al., 2019).
Fall detection and human activity recognition.
| Reference | AI Algorithm Best Achieved accuracy | Data Acquisition | Task |
|---|---|---|---|
|
| HMM with OpenPose | Two cameras | Fall risk assessment. Evaluation of imbalanced gait |
|
| SVM, 79% | micro-Doppler radar | Classification of gait differences associated with fall risk |
| CNN 73% | |||
|
| SOT, improved accuracy by 6% | Public HAR datasets UCI-DSADS, UCI-HAR, USC–HAD, PAMAP2 | Cross-domain HAR, utilizing transfer learning from auxiliary labeled data |
|
| NN | 11 men, TW, induced disturbances | Predict falls caused by an unexpected disturbance in time for CD to deploy |
|
| ANN, KNN, QSVM, EBT. fall detection = 100%, false alarms = 0, ARA = 97.7% | Wearable sensors Public datasets ( | ADL recognition. Fall detection |
|
| OpenPose for 2D pose estimation | Kinect images and sensor gait data from 250 subjects, 4 times, over 3 years | Estimation of Gait Parameters for Elderly Care from 3D Pose |
|
| RF, DT, KNN with K = 5. EER = 9.3 by RF. RF performs best in most of the sensor combinations | 51 subjects, 18 ADL. Smartphones in right pocket and smartwatch on the dominant hand | Continuous biometrics authentication and identification on smartphones or smartwatches. |
|
| SVM, KNN, NB, DT. Error 14.162% by SVM. | Inertial sensors. 19 subjects at home, 3 falls and 11 ADL | Wearable Fall Detection System |
|
| CSVD-NMF. 96.8% occupancy detection. 90.6% activity recognition | WiFi-enabled CSI measurements of 5 ADL | Device-Free Occupancy Sensing and activity recognition |
|
| Gaussian HMM. Sensitivity of 0.992. Positive predictive value of 0.981 | Own data. 200 fall events and 385 normal activities | Fall detection system |
|
| DCAE vs. CNN, SVM, AE. | micro-Doppler signatures | Radar-based activity recognition |
|
| ARA = 97.35% by GK-SVM. FD: sensitivity 98.70% and specificity 98.59% by GK-FDA. | 3 subjects, 7 ADL Wireless wearable sEMG sensors | Automatic activity recognition and fall detection |
|
| HCM-SFS on fused GRF and accelerometer data. ARA> 90% on all 5 ADL. | Force sensors and accelerometers under intelligent tiles. 6 subjects, 5 ADL | Fall detection and ADL recognition in independent living senior apartments |
|
| SVM, NN, DT, DA. 99% by SVM. | Smart phone IMU. 8 healthy subjects, 4 fall events, 6 ADL | ADL recognition and threshold-based fall detection |
|
| SVM | WiFi CSI measurements | Device-free wireless localization and activity recognition |
|
| Sparse BC+RVM. | 2 falling, 6 ADL, Spectrograms from continuous-wave radar | Radar-based Fall Detection |
| ( | KNN, kStar, HMM, SVM, DTC, RF, NB LR, ANN | smartphone | Activity recognition |
|
| SVM, KNN | inertial sensor | Recognition for similar gait action classes |
|
| k-means and KNN ANN + PCA | vision and sensor-based gait data | Abnormal gait detection |
|
| Variable-length PSO+ELM. 91.15% sensitivity, 77.14% specificity, and 86.83% accuracy | 10 young subjects, intentionally falling, and 6 ADL Kinect depth camera | Shape-based fall detection that is invariant to human translation, rotation, scaling and action length |
|
| KNN, LSM Over 99% | 14 subjects, 20 falls, 16 ADL, 6 wearable sensors | Automated fall detection system |
| ( | HMM | wireless IMU and an optical motion analysis system | Gait phase detection and walking/jogging discrimination |
Legend: Quadratic SVM (QSVM), HCM (Histogram Comparison Method), Sequential Forward Selection (SFS), Least squares method (LSM), Gaussian Kernel Fisher Discriminant Analysis (GK-FDA), Non-Negative Matrix Factorization (NMF), Class Estimated Basis Space Singular Value Decomposition (CSVD), Equal Error Rate (EER), Relevance Vector Machine (RVM), Gaussian Kernel SVM (GK-SVM), Substructural Optimal Transport (SOT), Channel State Information (CSI).
Datasets: UCI-DSADS (Anguita et al., 2012) UCI-HAR (Barshan and Yüksek, 2014), USC–HAD (Zhang and Sawchuk, 2012), PAMAP2 (Reiss and Stricker, 2012).
Tracking human pose.
| Reference | AI Algorithm Best Achieved accuracy | Data Acquisition | Purpose |
|---|---|---|---|
|
| Deep CNN. Error 21.54% | Low-power Radar. IDRad dataset made publicly available | Indoor PI invariant to the exact radar placement, room setup, and walking direction |
|
| Weka collection ML classifiers. 0.05 localization error. Accuracy > 93% | 4 Capacitive Sensors in load mode | Indoor Person Localization |
| (Li et al., 2015) | Improved PDR algorithm The best achieved accuracy is within 2 m | Samsung Galaxy Note3 and Bluetooth beacons | PDR algorithm integrated with Bluetooth beacons for indoor positioning without additional infrastructure |
|
| MagSLAM Achieves a position accuracy of 9–22 cm | Foot mounted IMU sensors. Low-power radar device. No a priori map | Dynamic positioning (SLAM) of indoor pedestrians derives a multi-floor indoor map |
Person identification (PI).
| Reference | Dataset/Input data | Proposed method for person identification |
|---|---|---|
| ( | CASIA-B and OU-MVLP | GaitPart: learns frame-level part spatial features and local short-range temporal features. Each body part has its own spatial-temporal representation |
|
| CASIA-B, OULP, and OUMVLP. | Proposed a gait-specific loss function called angle center loss. It uses learned horizontal partitions of gait templates and a temporal attention model |
|
| RSSI and phase features extracted from RF signals, 18 subjects | GRaaS; an RFID-based wireless gait recognition system using DRL tag selection algorithm and attention-based LSTM model |
|
| CASIA-B and OU-MVLP | GaitSet: Deep set-based PI using Set Pooling to aggregate silhouettes into one set |
| ( | OU-ISIR, CASIA-B, and USF | Multi-task GANs learn view-specific gait feature presentations. Proposed PEI, a new multi-channel gait template |
|
| SfootBD | Biometric Footstep Recognition using |
| Ensemble of ResNet and SVM with floor-only sensor data | ||
|
| VIPER, CUHK, and TownCentre | Crowd prototyping. Age, gender, and ethnicity recognition using ResNet-152 CNN |
| ( | TUM-GAID, CASIA-B, and OU-ISIR | Pose-based deep PI using WideResNet with OpenPose |
|
| CSI measurements using two routers in IoT network, 20 subjects | AutoID: WiFi-Based PI using C3SL |
|
| Radar micro-Doppler spectrograms, 24 subjects | RadarId: Deep CNN architecture based on raw radar micro-Doppler signatures |
|
| Constructed own IDRad dataset from FMCW radar, 5 subjects | Indoor PI using Deep CNN with radar data; PI in the dark, privacy-preserving, intruder detection |
|
| USF, CASIA-B, and OU-ISIR Own multi-gait image dataset from videos, 120 subjects in groups of 3 | A model-based method for multi-gait recognition using the L-CRF model |
|
| BIWI, IIT PAVIS, and IASLab | Depth-Based PI using a RNN/LSTM model. Suitable for PI in the dark |
|
| CASIA-B, PEC | Set-based PI using CNN, MLP, initialized with pretrained AlexNet |
|
| CASIA-B | Deep CNN framework for cross-view gait recognition, |
|
| McGill University and Osaka University gait datasets | KNN using GDI extracted from phone IMU data |
|
| USF and CASIA-B | A unitary ViDP, matrix projects the gait templates into a latent space for view-invariant PI |
|
| CASIA-B | 1-NN using proposed CGI as a gait template that preserves temporal information |
Legend: Convex Clustered Concurrent Shapelet Learning (C3SL), Latent Conditional Random Field (L-CRF), Gait Dynamics Image (GDI), View-Invariant Discriminative Projection (ViDP), Generative Adversarial Network (GAN), Period Energy Image (PEI), Radio Frequency Identification (RFID), Received Signal Strength Indicator (RSSI), Deep Reinforcement Learning (DRL), Gait Recognition as a Service (GRaaS), Frequency-Modulated Continuous-Wave (FMCW), ResNet-152 CNN (He et al., 2016), WideResNet (Zagoruyko and Komodakis, 2016).
Datasets: CASIA-B (Yu et al., 2006), OU-MVLP (Takemura et al., 2018), OU-ISIR (Iwama et al., 2012), OULP (Iwama et al., 2012), TUM-GAID (Hofmann et al., 2014), USF (Sarkar et al., 2005), PEC (Bossard et al., 2013), BIWI (Munaro et al., 2014), IIT PAVIS (Barbosa et al., 2012), and IASLab (Barbosa et al., 2012), McGill University Gait Dataset (Frank et al., 2010), Osaka University Gait Dataset (Ngo et al., 2014), SfootBD (Vera-Rodriguez et al., 2012), AlexNet (Krizhevsky et al., 2012).
Human Re-Identification.
| Reference | Dataset | Proposed method for ReID |
|---|---|---|
|
| Market1501, DukeMTMC-reID and CUHK03 | PrGCN; Graph based method. Predicts the link probability of the node pair |
|
| CASIA-B | PVTM: Transforms gallery image to the same view as the probe and uses only most informative human gait parts |
|
| iLIDS-VID, PRID 2011, and MARS. | Deep Siamese Attention Network Joint learning of spatiotemporal features and similarity metrics |
|
| PRID 2011, iLIDS-VID, and SDU-VID | Multiple CNN networks Compact appearance representation of selected frames rather than whole sequence |
|
| MARS and iLIDS-VID | D3DNet, Deep metric learning |
| Joint learning of spatiotemporal features and similarity metrics | ||
|
| VIPeR, CUHK01 | Stacked Auto-Encoders Deep metric learning of multiple similarity probabilities |
|
| MARS, Market-1501 and CUHK03 | MGCAM Binary segmentation mask and region-level triplet loss; Contrastive Learning |
|
| VIPeR, PRID 450S, and CUHK01 | Fine-tuned CNN with DM³. Matrix metric learning of discrepancy matrix instead of characteristic vector |
|
| ViPeR and CUHK01 | KNN, SVM ReID by saliency learning and matching |
|
| ViPeR, ETHZ, SAIVT-SoftBio, and iLIDS MCTS | Improved RDC, RankSVM and PCCA by using pose priors, image rectification and online person-specific weights |
|
| VIPeR, GRID, iLIDS MCTS, and CAVIAR4REID | RMLLC ReID as image retrieval task using relevance metric learning |
|
| VIPeR and ETHZ | MCE-KISS Improved KISS metric learning by MCE and a smoothing technique |
|
| GRID and VIPeR | MtMCML; multi-task learning. Designed multiple distance metrics |
|
| ETHZ, iLIDS MCTS, and VIPeR | Ensemble RDC model. Relative Distance Comparison Learning |
Legend: Probability Graph Convolutional Network (PrGCN), Dense 3D-Convolutional Network (D3DNet), Mask-guided Contrastive Attention Model (MGCAM), Discrepancy Matrix and Matrix Metric (DM³), Relevance Metric Learning with Listwise Constraints (RMLLC), Minimum Classification Error (MCE) Keep it simple and straightforward (KISS) Metric Learning, Multi-task Maximally Collapsing Metric Learning (MtMCML), Relative Distance Comparison (RDC), Support Vector Ranking (RankSVM), Pairwise Constrained Component Analysis (PCCA).
Datasets: VIPeR (Gray and Tao, 2008), CUHK01(Li et al., 2012), iLIDS-VID (Wang et al., 2014), PRID 2011 (Hirzer et al., 2011), and MARS (Zheng et al., 2016), SDU-VID (Liu et al., 2015), Market1501 (Felzenszwalb et al., 2008; Zheng et al., 2015), DukeMTMC-reID (Ristani et al., 2016), CUHK03 (Li et al., 2014), PRID 450S (Roth et al., 2014), GRID (Loy et al., 2009), iLIDS MCTS (Zheng et al., 2009), and CAVIAR4REID (Cheng et al., 2011), ETHZ (Schwartz and Davis, 2009), SAIVT-SoftBio (Bialkowski et al., 2012).
Person authentication (PA).
| Reference | AI Algorithm | Dataset/Data Modality | Purpose |
|---|---|---|---|
|
| SVDD and PCA for illegal user detection, LSTM for PI | Velocity and acceleration from the smartphone at the leg | PI and illegal user detection |
|
| Two-stream CNN with SVM | BrainRun dataaset. Own dataset of gait and other behavioral features from smartphones, 100 subjects. | SCANet: Continuous PA, distinguishes legitimate vs impostor users |
| ( | Multi-layer LSTM and Extreme Value Statistic | ZJU-GaitAcc, 3D accelerations from smartphones | PI and PA of the learned user, reject unauthorized user |
| ( | SVM, KNN, DT | acceleration, angular velocity, magnetic intensity, and PPG signals from fingertip device | Multisensor PA, HAR |
|
| Deep CNN | Own IDRad Dataset: micro-Doppler signatures, 5 subjects | Automatic intruder detection, indoor PI |
|
| Semi-supervised ML, Isolation Forest | Tracking current vs. known usage of the device and motion sensor data from phone | Adaptive and continuous PA system, anomaly detection |
|
| Dense clockwork RNN | HMOG, Google Abacus Dataset: time series of inertial measurements | distinguishes legitimate vs impostor users |
Legend: Photoplethysmography (PPG), Support Vector Data Description (SVDD), Growing Neural Gas (GNG).
Datasets: BrainRun (Papamichail et al., 2019), ZJU-GaitAcc (Zhang Y. et al., 2014), HMOG (Yang et al., 2014), UMN (Raghavendra et al., 2006), UCSD Ped (Li et al., 2013), Avenue (Lu et al., 2013).
Gender recognition (GR).
| Reference | AI Algorithm Best Achieved accuracy | Dataset/Input features | Task |
|---|---|---|---|
|
| KNN, SVM, NB, DT, 100% | UPCVgaitK1, UPCVgaitK2 | GR |
|
| multi-task CNN, AE: MAE = 5.47, GR: 98.1% | OULP-Age dataset GEI from video | GR and AE |
|
| LK-SVM with FLBP Normal: 96.40% | CASIA-B GEI from video | GR |
| Carrying: 87.97% | |||
| Wearing coat: 86.54% | |||
| ( | Bootstrap DT 94.44% | 1D HG extracted from Smartphone in the front pocket | GR |
| ( | CNN F:77%, M:96% | TUM-GAID: extracted from low-resolution video streams recorded with MS Kinect | automatic PI and GR |
|
| AP clustering + SRML PI: 87.6% GR: 93.1% | Own dataset: ADSCAWD USF and CASIA-B C-AGI instead of GEI from MS Kinect Depth Sensor | PI and GR |
Legend: Sparse Reconstruction-based Metric Learning (SRML), Cluster-based Averaged Gait Image (C-AGI), Affinity Propagation (AP), Optical Flow (OF), Person Identification (PI), Gender Recognition (GR), Age Estimation (AE), Fuzzy Local Binary Pattern (FLBP), Linear Kernel SVM (LK-SVM).
Datasets: UPCVgaitK1 (Kastaniotis et al., 2013), UPCVgaitK2 (Kastaniotis et al., 2016), OULP-Age (Iwama et al., 2012), CASIA-B (Yu et al., 2006), TUM-GAID (Hofmann et al., 2014), USF (Sarkar et al., 2005).
Smart gait devices.
| Reference | AI Algorithm | AI task | Sensing Technology | Application |
|---|---|---|---|---|
|
| LSTM with CAE | estimating joint torque for motion intent prediction | three soft pneumatic sensors two 3D IMUs | soft smart shoes |
|
| attention-based LSTM | gait recognition while preserving privacy of users | KEH | PrivGait, a KEH-equipped wearable device |
|
| RF | human activity recognition | wrist-worn solar cell | SolAR, a solar self-powered wearable device |
|
| 1D CNN | gait and human activity recognition | textile TENGs | smart socks for long-term gait monitoring |
|
| NB, RF, DT, KNN | human activity recognition | two capacitors and two transducers | a self-powered shoe with embedded CapSense technology |
|
| RNN | imitation learning for real-time prosthetic control | built-in motion sensors | powered transfemoral prosthesis |
|
| RL | assist-as-needed control for robot-assisted gait training | built-in motion sensors | SAFE orthosis |
|
| DDNN | classification of EMG signals to activate an event-driven controller | EMG sensors | mobile lower limb active orthosis |
Legend: Deep Differential Neural Networks (DNNN), Stevens Ankle-Foot Electromechanical (SAFE), Reinforcement Learning (RL).
Smart home applications.
| Reference | AI Algorithm | Data Acquisition | Purpose |
|---|---|---|---|
|
| AI algorithms are built into smart objects | Wall light for indoor localization. | This paper provides a complete description of the HABITAT project regarding methodology, architecture, design, and smart objects development |
| The smart armchair and the smart belt perform activity recognition algorithms. | Armchair for sitting posture monitoring.The belt for movement information. | ||
| The wall light sends input to a fall detection algorithm | The Wall panel and mobile devices are the user interface | ||
|
| 3D gesture recognition: 95.3% using PNN and 10-fold CV.Pedestrian navigation: distance and positioning accuracies were 0.22 and 3.36% of the total distance traveled in the indoor environment.Home safety and fire detection: classification rate 98.81%. | Wearable IMU on wrist tracks hand gesture and on feet walking data and energy managementEnvironmental sensors.Experimental smart home testbed. Web camera. Multisensory circuit module for home safety and fire detection | Design and implementation of a smart home system that integrates wearable intelligent technology, artificial intelligence, and sensor fusion technology to complete these tasks:Automated household appliance control.Smart energy management. |
| PNN, DTW, SVM, LDA, PDR, PCA-PNN. | Fire detection and home safety | ||
|
| DT, NBC, RF, SVM, Ada/DT, Ada/RF are tried out, and Ada/DT provides the best classification accuracy. | CASAS smart home and wearable sensors.Analysis 1: | In home health monitoring for early detection of changes associated with PD and MCI and evaluation of treatment |
| PCA is used to reduce features k-Means clustering and random resampling are used to add features in smaller (individual activities) datasets | Analysis 2: |
Legend: Pseudo-odometry (P-O), Adaptive Boosting (Ada), Probabilistic Neural Network (PNN), Mild Cognitive Impairment (MCI), Instrumental ADLs (IADLs), Dynamic Bayesian Network (DBN).
Animation and virtual environments.
| Reference | AI Algorithm/Characteristics | Data Acquisition/Inputs | Task |
|---|---|---|---|
|
| THR, COR, SVM and BiLSTM, tested |
| Motion reconstruction |
| - COR has the best accuracy for real-time VR applications (low delay) | Gait phase detection | ||
|
| DReCon: motion matching and deep RL | Unstructured motion data from mocap | Real-time physics-based character control for video games |
| - responsive to user demands, natural-looking. Trained on flat terrain | |||
|
| OpenPose/HMR and DRL | Simulated character model and YouTube video clip | Learning dynamic physics-based character controllers from video clips |
| - Learning from inexpensive video clips, robust | |||
|
| DeepMimic: DRL | Character model, kinematic reference motion from video clip | Physics-based character controllers from video clips |
| - Diverse skills/terrains/morphologies, realistic response to perturbations | |||
|
| SMPL body model and BiLSTM | 6 IMUs | 3D human pose reconstruction from a sparse set of IMUs |
| - Useful when camera-based data is not available due to occlusion, fast motion, etc | |||
|
| CAE | CMU Motion Capture Database ³ | Unsupervised learning of a human motion manifold |
| - Capable of fixing corrupt data, filling in missing data, motion interpolation along the manifold, and motion comparison | |||
|
| SMG and part-based Laplacian deformation | Three 4DPC datasets 4 | A data-driven approach for animating 4DPC character models |
| - Simultaneously captures both motion and appearance for video-like quality | |||
|
| Multilayer JGPMs/topologically constrained GPLVMs | CMU Motion Capture Database + Simulated data | Human gait modeling |
| - diversity of walking styles, motion interpolation, reconstruction, and filtering | |||
|
| FFSM with automatic learning of the fuzzy KB by GA |
| Human gait modeling |
| - Fuzzy states and transitions are still defined by experts, interpretable, generalizes well for each person’s gait | Accelerometer attached to the belt |
Legend: Threshold Based Method (THR), Pearson Correlation-based Method (COR), Data-Driven Responsive Control (DReCon), Human Mesh Recovery (HMR), Deep Deterministic Policy Gradient (DDPG), Skinned Multi-Person Linear (SMPL) as in (Loper et al., 2015), 4D Performance Capture (4DPC), Surface Motion Graphs (SMGs), Carnegie Mellon University (CMU), Joint Gait-Pose Manifolds (JGPMs), Fuzzy Finite State Machines (FFSM), Knowledge Base (KB).
Datasets: ³ http://mocap.cs.cmu.edu/ 4 http://cvssp.org/cvssp3d.