| Literature DB >> 31137677 |
Hamid Heydarian1, Marc Adam2,3, Tracy Burrows4,5, Clare Collins6,7, Megan E Rollo8,9.
Abstract
Wearable motion tracking sensors are now widely used to monitor physical activity, and have recently gained more attention in dietary monitoring research. The aim of this review is to synthesise research to date that utilises upper limb motion tracking sensors, either individually or in combination with other technologies (e.g., cameras, microphones), to objectively assess eating behaviour. Eleven electronic databases were searched in January 2019, and 653 distinct records were obtained. Including 10 studies found in backward and forward searches, a total of 69 studies met the inclusion criteria, with 28 published since 2017. Fifty studies were conducted exclusively in laboratory settings, 13 exclusively in free-living settings, and three in both settings. The most commonly used motion sensor was an accelerometer (64) worn on the wrist (60) or lower arm (5), while in most studies (45), accelerometers were used in combination with gyroscopes. Twenty-six studies used commercial-grade smartwatches or fitness bands, 11 used professional grade devices, and 32 used standalone sensor chipsets. The most used machine learning approaches were Support Vector Machine (SVM, n = 21), Random Forest (n = 19), Decision Tree (n = 16), Hidden Markov Model (HMM, n = 10) algorithms, and from 2017 Deep Learning (n = 5). While comparisons of the detection models are not valid due to the use of different datasets, the models that consider the sequential context of data across time, such as HMM and Deep Learning, show promising results for eating activity detection. We discuss opportunities for future research and emerging applications in the context of dietary assessment and monitoring.Entities:
Keywords: accelerometer; eating activity detection; gyroscope; hand-to-mouth movement; wrist-mounted motion tracking sensor
Mesh:
Year: 2019 PMID: 31137677 PMCID: PMC6566929 DOI: 10.3390/nu11051168
Source DB: PubMed Journal: Nutrients ISSN: 2072-6643 Impact factor: 5.717
Terms used in this review with synonyms and definitions.
| Term | Synonyms Used in the Literature | Definition |
|---|---|---|
|
| Event/activity classifications/categories | Different categories that the classifiers (detection models) are trained and tested to classify |
|
| Artificial intelligence approach, machine learning algorithm | The approach used for automatic eating behaviour detection |
|
| Search through references of included studies to find other relevant studies that are not found through database search | |
|
| Eating and drinking activity | |
|
| Food intake detection, eating detection, ingestion monitoring | Assessing whether the participant is eating (including drinking) and what their eating characteristics are |
|
| Search for relevant studies that cited included studies | |
|
| F1 score, F-measure | F-score is a measure of accuracy. While accuracy is the total number of correctly classified items divided by all classified items, F-score is harmonic average of the precision and recall. |
|
| Hand-to-mouth movement | The movement of hand carrying food with or without utensils to the mouth |
|
| Motion tracking sensors, motion detection sensors, activity tracker | Sensors used to detect movements. Wearable motion sensors focused on in the current review include upper limb-mounted motion sensors. |
|
| Subject | An individual who has successfully participated in a study (i.e., not counting individuals who were invited but did not participate or individuals with failed measurements) |
|
| Arm | Region of body that includes shoulder, upper arm, lower arm, wrist, and hand |
Search strategy string.
| Search String |
|---|
| (accelerometer OR gyroscope OR smartwatch OR “inertial sensor” OR “inertial sensors” OR “inertial sensing” OR smartphone OR “cell phone” OR wristband) AND (“dietary intake” OR “dietary assessment” OR “food intake” OR “nutrition assessment” OR “eating activity” OR “eating activities” OR “eating behavior” OR “eating behaviour” OR “energy intake” OR “detecting eating” OR “detect eating” OR “eating episodes” OR “eating period”) AND (“bite counting” OR “counting bites” OR “hand gesture” OR “hand gestures” OR “arm gesture” OR “arm gestures” OR “wrist gesture” OR “wrist gestures” OR “hand motion” OR “hand motions” OR “arm motion” OR “arm motions” OR “wrist motion” OR “wrist motions” OR “hand movement” OR “hand movements” OR “arm movement” OR “arm movements” OR “wrist movement” OR “wrist movements” OR “hand to mouth” OR “hand-to-mouth” OR “wrist-worn” OR “wrist-mounted”) |
Search strategy databases (English only results).
| Database name | Result |
|---|---|
| ACM | 10 |
| AIS Electronic Library | 55 |
| CINAHL | 16 |
| EMBASE 1 | |
| IEEE 2 | 140 |
| MEDLINE 1 | |
| Ovid databases 3 | 54 |
| ScienceDirect 4 | 133 |
| Scopus | 161 |
| SpringerLink | 205 |
| Web of Science | 18 |
| Total results before removing duplicates | 792 |
| Total results after removing duplicates | 653 |
1 Results from Ovid databases include results from databases EMBASE and MEDLINE. 2 Due to limitation on the length of the search string that IEEE database accepted, the search string was broken up to smaller parts (see Supplementary Material). 3 Ovid databases include Books@Ovid, Embase, Emcare, MEDLINE. The following option were chosen to be included in the results: AMED (Allied and Complementary Medicine) 1985 to September 2017, Books@Ovid September 25, 2017, Embase 1947 to present, Emcare (Nursing and Allied Health) 1995–present, International Pharmaceutical Abstracts 1970 to September 2017, Medline 1946–present, University of Newcastle Journals. 4 ScienceDirect database web search did not accept the search terms “hand-to-mouth” and “hand to mouth”. Therefore, these two search terms were excluded from the search string submitted to ScienceDirect. Also due to limitation on the length of the search string that ScienceDirect database accepted, the search string was broken up to smaller parts (see Supplementary Material).
Included studies (n = 69).
| Author (Year), Country [Ref] | Type | Participants (Environment) | Upper Limb Sensor (Frequency) [Position] | Comparator | Algorithm (Detection) |
|---|---|---|---|---|---|
| Amft et al. (2005), Switzerland [ | Conference | 2 (lab) | Acc/Gyro (100 Hz) [wrist, both] | NR | FSS, HMM (GD) |
| Amft et al. (2007), Switzerland [ | Conference | 1 (lab) | Acc/Gyro (NR) [lower arm, both] | NR | PCFG (AD) |
| Amft et al. (2008), Switzerland [ | Journal | 4 (lab) | Acc/Gyro (100 Hz) [lower arm, both] | NR | FSS (GD) |
| Junker et al. (2008), Switzerland [ | Journal | 4 (Lab) | Acc/Gyro (100 Hz) [lower arm, both] | NR | HMM (GD) |
| Laerhoven et al. (2008), Germany [ | Conference | 2 (free-living) | Acc (NR) [wrist, dominant] | Other self-report | KNN (AD) |
| Pirkl et al. (2008), Germany [ | Conference | 1 (lab) | Acc/Gyro (50 Hz), Prox [lower arm, dominant] | NR | DT (GD) |
| Tolstikov et al. (2008), Singapore [ | Conference | NR (lab) | Acc (50 Hz), Prox [wrist, dominant] | Camera | DBN (AD) |
| Zhang et al. (2008), Singapore [ | Conference | NR (lab) | Acc (NR) [wrist, both] | NR | HTM (GD) |
| Dong et al. (2009), USA [ | Conference | 10 (lab) | Acc/Gyro (60 Hz) [wrist, dominant] | Camera | C/RB (GD) |
| Teixeira et al. (2009), USA [ | Conference | 1 (free-living) | Acc/Gyro (NR) [wrist, dominant] | NR | FSM (AD) |
| Amft et al. (2010), Netherlands [ | Conference | 9 (lab) | Acc/Gyro (87 Hz), Prox [wrist, dominant] | Camera | DT, FSS (GD) |
| Dong et al. (2011), USA [ | Conference | 4 (free-living) | Acc/Gyro (60 Hz) [wrist, dominant] | Diary | C/RB (AD) |
| Dong et al. (2012), USA [ | Journal | 102 (both) | Acc/Gyro (NR) [wrist, dominant] | Camera, Diary | C/RB (GD) |
| Grosse-Puppendahl et al. (2012), Germany [ | Chapter | 7 (lab) | Acc (NR), Prox (NR) [wrist, dominant] | NR | SVM (AD) |
| Kim et al. (2012), South Korea [ | Conference | 13 (lab) | Acc (30 Hz) [wrist, dominant] | Camera | DT (CD, GD) |
| Mousavi Hondori et al. (2012), USA [ | Conference | 1 (lab) | Acc (NR) [utensil, both] | NR | NR (NR) |
| Varkey et al. (2012), USA [ | Journal | 1 (lab) | Acc/Gyro (20 Hz) [wrist, dominant] | NR | SVM (AD) |
| Farooq et al. (2013), USA [ | Conference | 13 (free-living) | Prox (NR) [wrist, dominant] | Diary, Push Button | ANN, SVM (AD) |
| Fontana et al. (2013), USA [ | Conference | 12 (free-living) | Prox (10 Hz) [wrist, dominant] | Push Button | RF (AD) |
| Kim & Choi (2013), South Korea [ | Conference | 8 (lab) | Acc (30 Hz) [wrist, dominant] | Camera | DT, NB (AD, CD, GD) |
| Ramos-Garcia & Hoover (2013), USA [ | Conference | 273 (lab) | Acc/Gyro (15 Hz) [wrist, dominant] | Camera | HMM, KNN (GD) |
| Desendorf et al. (2014), USA [ | Journal | 15 (lab) | Acc (80 Hz) [wrist, dominant] | NR | C/RB (GD) |
| Dong et al. (2014), USA [ | Journal | 43 (free-living) | Acc/Gyro (NR) [lower arm, dominant] | Mobile App | C/RB, NB (GD, AD) |
| Fontana et al. (2014), USA [ | Journal | 12 (free-living) | Prox (NR) [wrist, dominant] | Mobile App, Push Button | ANN (AD) |
| Ramos-Garcia et al. (2015), USA [ | Journal | 25 (lab) | Acc/Gyro (15 Hz) [wrist, dominant] | Camera | HMM, KNN (GD) |
| Sen et al. (2015), Singapore [ | Conference | 6 (lab) | Acc/Gyro (100 Hz) [wrist, dominant] | Wearable Camera | C/RB (GD) |
| Thomaz et al. (2015), USA [ | Conference | 28 (both) | Acc (25 Hz) [wrist, dominant] | Camera, Wearable Camera | DBSCAN, KNN, RF, SVM (AD, GD) |
| Ye et al. (2015), USA [ | Conference | 10 (lab) | Acc (50 Hz) [wrist, dominant] | Camera | DT, NB, SVM (GD) |
| Zhou et al. (2015), Japan [ | Journal | 5 (lab) | Acc (NR) [finger, dominant] | NR | DT, KNN (GAD) |
| Fan et al. (2016), USA [ | Journal | 1 (lab) | Acc/Gyro (50.1 Hz) [wrist, dominant] | NR | ANN, DT, KNN, NB, Reg, RF, SVM (AD) |
| Farooq & Sazonov (2016), USA [ | Conference | 12 (free-living) | Prox. (NR) [wrist, dominant] | Mobile App, Push Button | DT, Reg (AD) |
| Fortuna et al. (2016), USA [ | Conference | 3 (free-living) | Acc/Gyro (10 Hz) [wrist, dominant] | Camera, Wearable Camera | NB (GD) |
| Kim et al. (2016), South Korea [ | Conference | 15 (lab) | Acc/Gyro (NR) [wrist, dominant] | Camera | C/RB (GD) |
| Maramis et al. (2016), Greece [ | Conference | 8 (lab) | Acc/Gyro (NR) [wrist, dominant] | Camera | SVM (GD) |
| Mirtchouk M. et al. (2016), USA [ | Conference | 6 (lab) | Acc/Gyro (15 Hz) [wrist, both] | Camera | RF (CD) |
| Parra-Sanchez et al. (2016), Mexico [ | Conference | 7 (lab) | Electro-hydraulic (NR) [arm, both] | Camera | DT, HMM (GAD) |
| Rahman et al. (2016), Canada [ | Conference | 8 (free-living) | Acc/Gyro (15 Hz) [wrist, dominant] | Mobile App | DT, Reg, RF, SVM (AD, CD) |
| Sharma et al. (2016), USA [ | Conference | 94 (free-living) | Acc/Gyro (15 Hz) [wrist, dominant] | NR | C/RB, NB (AD, GD) |
| Shen et al. (2016), USA [ | Conference | 215 (lab) | Acc/Gyro (15 Hz) [wrist, dominant] | Camera | HMM (GD) |
| Shoaib et al. (2016), Netherlands [ | Conference | 11 (lab) | Acc/Gyro (50 Hz) [wrist, dominant] | NR | DT, RF, SVM (AD) |
| Ye et al. (2016), USA [ | Conference | 7 (free-living) | Acc (50 Hz) [wrist, dominant] | Mobile App | SVM (GD) |
| Zhang et al. (2016), USA [ | Conference | 15 (lab) | Acc/Gyro (31 Hz) [wrist, both] | Camera | DBSCAN, DT, NB, Reg, RF, SVM (GD, AD) |
| Alexander et al. (2017), USA [ | Journal | 4 (lab) | Acc/Gyro (20 Hz) [wrist, dominant] | Time Sync | NB (AD) |
| Bi et al. (2017), USA [ | Journal | 37 (free-living) | Acc (80 Hz) [wrist, dominant] | Other self-report | HMM, SVM (AD) |
| Dong & Biswas (2017), USA [ | Journal | 14 (lab) | Acc (100 Hz) [wrist, dominant] | Camera, Push Button | C/RB, HMM, SVM (GD, AD) |
| Egilmez et al. (2017), USA [ | Journal | 9 (lab) | Acc/Gyro (5 Hz) [wrist, dominant] | NR | NB, Reg, RF, SVM (GAD) |
| Garcia-Ceja et al. (2017), Mexico [ | Journal | 3 (lab) | Acc/Gyro (31 Hz) [wrist, dominant] | NR | RF (GAD) |
| Kyritsis et al. (2017), Greece [ | Conference | 8 (lab) | Acc/Gyro (62 Hz) [wrist, dominant] | Camera | HMM, SVM (AD, GD) |
| Kyritsis et al. (2017), Greece [ | Conference | 10 (lab) | Acc/Gyro (62 Hz) [wrist, dominant] | Camera | DL, SVM (AD, GD) |
| Loke & Abkenar (2017), Australia [ | Journal | 4 (lab) | Acc (NR) [wrist, dominant] | NR | DT (GAD) |
| Moschetti et al. (2017), Italy [ | Conference | 12 (lab) | Acc/Gyro (50 Hz) [wrist, dominant] | NR | GMM, KM, RF, SVM (GD) |
| Sen et al. (2017), Singapore [ | Journal | 28 (both) | Acc/Gyro (NR) [wrist, dominant] | Camera, Diary | DT, RF, SVM (AD, GD) |
| Shen et al. (2017), USA [ | Journal | 271 (lab) | Acc/Gyro (15 Hz) [wrist, dominant] | Camera | C/RB (GD) |
| Soubam et al. (2017), India [ | Conference | 11 (lab) | Acc (186 Hz) [wrist, dominant] | Time Sync | DT, NB, RF, SVM (CD, GD) |
| Thomaz et al. (2017), USA [ | Conference | 14 (lab) | Acc/Gyro (30 Hz) [wrist, both] | Camera | RF (AD) |
| Yoneda & Weiss (2017), USA [ | Conference | 51 (lab) | Acc/Gyro (20 Hz) [wrist, dominant] | NR | DT, KNN, RF (GAD) |
| Zhang et al. (2017), USA [ | Conference | 8 (free-living) | Acc/Gyro (31 Hz) [wrist, both] | Wearable Camera | RF (GD) |
| Anderez et al. (2018), UK [ | Conference | NR (lab) | Acc/Gyro (100 Hz) [wrist, dominant] | NR | DL, RF (GD) |
| Anderez et al. (2018), UK [ | Conference | NR (lab) | Acc (100 Hz) [wrist, dominant] | NR | KNN (GD) |
| Balaji et al. (2018), India [ | Conference | NR (NR) | Acc/Gyro (100 Hz) [wrist, NR] | NR | C/RB (GAD) |
| Cho & Choi (2018), South Korea [ | Conference | 8 (lab) | Acc (50 Hz) [wrist, dominant] | Camera | DL (CD, GD) |
| Clapés et al. (2018), Spain [ | Journal | 14 (lab) | Acc/Gyro (25 Hz) [wrist, dominant] | Camera | ANN, Opt (GAD, GD) |
| Kyritsis et al. (2018), Greece [ | Conference | 10 (lab) | Acc/Gyro (100 Hz) [wrist, dominant] | Camera | DL (AD) |
| Manzi et al. (2018), Italy [ | Journal | 20 (lab) | Acc/Gyro (NR) [wrist, dominant] | NR | RF (GAD) |
| Papadopoulos et al. (2018), Greece [ | Conference | 10 (lab) | Acc/Gyro (62 Hz) [wrist, dominant] | Camera | semi-supervised DL (AD) |
| Schibon & Amft (2018), Germany [ | Conference | 6 (free-living) | Acc/Gyro (NR) [wrist, both] | Diary | SVM (GD) |
| Shen et al. (2018), USA [ | arXiv | 269 (lab) | Acc/Gyro (15 Hz) [wrist, dominant] | Camera | HMM (GD) |
| Zambrana et al. (2018), Spain [ | Journal | 21 (lab) | Acc (20 Hz) [wrist, both] | Camera | KNN, RF, SVM (GAD) |
| Zhang et al. (2018), USA [ | Journal | 10 (lab) | Acc/Gyro (NR) [wrist, dominant] | Camera | RF (GD) |
The table only includes performance comparisons where studies have directly compared different algorithms using the same dataset. Acc = Accelerometer, AD = Eating Activity Detection, ANN = Artificial Neural Network, CD = Eating Characteristics Detection, Chap = Book Chapter, Conf = Conference, C/RB = Custom Rule-Based, DBN = Dynamic Bayesian Network, DBSCAN = Density-Based Spatial Clustering of Applications with Noise, DL = Deep Learning, DT = Decision Tree, FL = Free-Living, FSM = Finite State Machine, FSS = Feature Similarity Detection, GAD = General Activity Detection, GD = Eating Gesture Detection, GMM = Gaussian Mixture Model, Gyro = Gyroscope, HMM = Hidden Markov Model, HTM = Hierarchical Temporal Memory, Jour = Journal, KM = K-Means, KNN = K-Nearest Neighbours, NB = Naive Bayes, NR = Not Reported, Opt = Monte Carlo Optimization method, PCFG = Probabilistic Context-Free Grammar, Prox = Proximity, Reg = Regression, RF = Random Forest, SVM = Support Vector Machine.
Figure 1Flow diagram of article selection process in the systematic review.
Figure 2Conceptual framework of components for assessing eating behaviour with upper limb-mounted motion sensors.
Machine learning algorithms used in the included studies as well as and detection approaches and performance comparisons conducted in the studies.
| Algorithm (# and % Studies)—Approach | Best Performing Algorithm (vs. Comparison Algorithm/s) | Performance Comparison Results |
|---|---|---|
| SVM (vs. DT, NB) [ | Using only wrist or head motion data SVM accuracy for eating detection was between 0.895 to 0.951 (combined wrist/head: 0.970). | |
| SVM (vs. KNN, RF) [ | Best accuracy of SVM using time-domain features was 0.957 (F = 0.957, two-second window). Best accuracy of RF model using frequency-domain features was 0.939 (F = 0.940, four-second window). | |
| SVM (vs. DT, RF) [ | F-scores for detecting eating, drinking, and smoking with SVM were 0.910, 0.780, and 0.830, compared to RF with 0.860, 0.780, and 0.840, and DT with 0.820, 0.690, and 0.780. | |
| RF (vs. SVM, KNN) [ | FR outperformed SVM and 3-NN in detecting eating gestures in two free-living settings (seven participants in one day, F = 0.761; one participant in 31 days (F = 0.713). | |
| RF (vs. DT, NB, Reg, SVM) [ | Eating gesture detection with RF yielded F = 0.753 compared to F = 0.708 (SVM), F = 0.694 (Reg), F = 0.647 (DT), and F = 0.634 (NB). | |
| RF (vs. DT, SVM) [ | The accuracies achieved by RF, DT, and SVM were 0.982, 0.966, and 0.857, respectively. | |
| RF (vs. SVM) [ | Using leave one person out cross-validation method the accuracies achieved by RF and SVM were 0.943 (F = 0.949) and 0.882 (F = 0.895). | |
| RF (vs. NB, Reg, SVM) [ | F-scores for general activity detection with FR, Reg, SVM, and NB were 0.788, 0.661, 0.6l3, and 0.268, respectively (eating was among action classes). | |
| RF (vs. DT, NB, SVM) [ | The accuracies of RF, SVM, DT, and NB in person independent drinking versus eating detection were 0.924, 0.905, 0.881 and 0.871, respectively. | |
| RF (vs. DT, Reg, SVM) [ | Person independent “about-to-eat” detection achieved F = 0.690 (RF) compared to 0.660 (SVM), 0.660 (REG), and 0.640 (DT). | |
| RF (vs. DT, KNN) [ | Accuracies of RF, DT, and KNN were 0.997, 0.980, and 0.888, respectively (using accelerometer data). | |
| DT (vs. NB) [ | Best F-scores of DT and NB were 0.930 and 0.900 (eating activity detection) and 0.780 and 0.700 (eating type detection), respectively. | |
| DT (vs. NB) [ | Best F-scores of DT and NB were 0.750 and 0.650 (eating utensil detection) and 0.280 and 0.190 (eating action detection), respectively. | |
| HMM (vs. SVM) [ | HMM outperformed SVM by 6.82% on recall in family meal detection, while the average precision and recall were 0.807 and 0.895. | |
| HMM (vs KNN) [ | Accuracies of HMM, and KNN were 0.843 and 0.717, respectively. | |
| HMM-1 (vs. HMM-S, HMM-N, N:2-6) [ | Accuracies of HMM-S and gesture-to-gesture HMM-1 were 0.852 and 0.895 respectively. According to the figure provided in the study the accuracy for HMM-2 to 4 stays similar and decreases for HMM-5 and 6. | |
| HMM-6 (vs. HMM-N, N:1-5, KNN, S-HMM) [ | The accuracies of HMM-6, HMM-5, HMM-4, HMM-3, HMM-2, HMM-1, sub-gesture HMM and KNN were 0.965, 0.946, 0.922, 0.896, 0.880, 0.877, 0.843 and 0.758, respectively. | |
| KNN (vs. ANN, DT, NB, Reg, RF, SVM) [ | The accuracies of KNN, RF, Reg, SVM, ANN, NB and DT were 0.936, 0.933, 0.923, 0.920, 0.913, 0.906 and 0.893, respectively. | |
| KNN (vs. DT, NB) [ | The precision (and recall) values of KNN, DT and NB were 0.710 (0.719), 0.670 (0.686) and 0.657 (0.635), respectively. | |
| RNN (vs. HMM) [ | Replacing HMM with RNN in a SVM-HMM model improved F-score from 0.814 to 0.892 [ | |
| ANN (vs. SVM) [ | ANN achieved accuracy of 0.869 (±0.065) compared to SVM (0.819, ±0.092) for eating activity detection (12 participants). ANN achieved accuracy of 0.727 compared to SVM (0.636, ±0.092) for number of meals detection (1 participant). | |
| KM (vs. GMM) [ | In an inter-person comparison, the accuracies of unsupervised approaches KM and GMM were 0.917 (F = 0.920) and 0.796 (F = 0.805), respectively. | |
AD = Eating Activity Detection, ANN = Artificial Neural Network, CD = Eating Characteristics Detection, C/RB = Custom Rule-Based, DBN = Dynamic Bayesian Network, DBSCAN = Density-Based Spatial Clustering of Applications with Noise, DL = Deep Learning, DT = Decision Tree, F = F-score, FSM = Finite State Machine, FSS = Feature Similarity Detection, GAD = General Activity Detection, GD = Eating Gesture Detection, GMM = Gaussian Mixture Model, HMM = Hidden Markov Model, HMM-S = single-gesture HMM, HTM = Hierarchical Temporal Memory, KM = K-Means, KNN = K-Nearest Neighbours, NB = Naive Bayes, Opt = Monte Carlo Optimization method, PCFG = Probabilistic Context-Free Grammar, Reg = Regression, RF = Random Forest, RNN = Recurrent Neural Network, SVM = Support Vector Machine.