| Literature DB >> 33592164 |
Han Shi Jocelyn Chew1, Wei How Darryl Ang1, Ying Lau1.
Abstract
OBJECTIVE: To present an overview of how artificial intelligence (AI) could be used to regulate eating and dietary behaviours, exercise behaviours and weight loss.Entities:
Keywords: Artificial intelligence; Behaviour change; Diet; Obesity; Self-control; Self-regulation; Weight
Year: 2021 PMID: 33592164 PMCID: PMC8145469 DOI: 10.1017/S1368980021000598
Source DB: PubMed Journal: Public Health Nutr ISSN: 1368-9800 Impact factor: 4.022
Fig. 1PRISMA 2009 flow diagram for first search
Fig. 2PRISMA 2009 flow diagram for second search
Fig. 3Data mapping of AI features used for different self-regulation components (n 66)
Study characteristics (n 66)*
| Study characteristics |
| % |
|---|---|---|
| Country of study | ||
| USA | 37 | 56·1 |
| Canada | 3 | 4·5 |
| Finland | 2 | 3·0 |
| France | 1 | 1·5 |
| Germany | 1 | 1·5 |
| Greece | 1 | 1·5 |
| Switzerland | 2 | 3·0 |
| The Netherlands | 1 | 1·5 |
| UK | 5 | 7·6 |
| China (including one study from Hong Kong) | 6 | 9·1 |
| South Korea | 2 | 3·0 |
| Taiwan | 2 | 3·0 |
| India | 1 | 1·5 |
| Saudi Arabia | 1 | 1·5 |
| Australia | 1 | 1·5 |
| Study design | ||
| RCT | 8 | 12·1 |
| Experimental (lab-based, quasi-experimental, pretest-posttest) | 50 | 75·8 |
| Observational | 6 | 9·1 |
| Secondary data analysis | 2 | 3·0 |
| Sample size | ||
| < 100 | 54 | 81·8 |
| 100–300 | 5 | 7·6 |
| Not reported | 8 | 12·1 |
| Sex | ||
| All females | 1 | 1·5 |
| All males | 0 | 0 |
| Both sexes | 59 | 89·4 |
| Not reported | 7 | 24 |
| Mean age (years old): | ||
| 20–30 | 19 | 28·8 |
| 31–40 | 7 | 10·6 |
| 41–50 | 3 | 4·5 |
| 50–60 | 5 | 7·6 |
| Not reported | 28 | 42·4 |
| Baseline BMI (kg/m2) | ||
| Reported | 37 | 56·1 |
| Not reported | 30 | 45·5 |
One included article consisted of two studies; hence, the total number of studies is 66.
Includes ten studies that provided age ranges.
Functions of AI in self-regulation of weight management in healthy and overweight populations (n 66)
| Author, year | Self-regulation use case | AI features | AI functions | Data collection instrument | Data type | Important results |
|---|---|---|---|---|---|---|
| Alshurafa | Self-monitoring (eating behaviour) | Gesture recognition | To distinguish between food types | Necklace piezoelectric sensor (vibration sensor) | Skin motion when swallowing | Successfully distinguished between liquids and solids ( |
| Amft | Self-monitoring (eating behaviour) | Gesture and sound recognition | To recognise dietary activities using on-body sensors | On-body sensors: | Arm movements, chewing cycle sounds and swallowing | Four intake gestures from arm movements and two food groups from chewing cycle sounds were detected and identified with a recall of 80–90 % and a precision of 50–64 %. The detection of individual swallows resulted in 68 % recall and 20 % precision. Sample-accurate recognition rates were 79 % for movements, 86 % for chewing and 70 % for swallowing |
| Amft | Self-monitoring (eating behaviour) | Sound recognition | To evaluate the prediction of food weight in individual bites using an ear-pad chewing sound sensor | Ear-pad chewing sound sensor | Chewing cycles and food type | Sound-based chewing recognition achieved recalls of 80 % at 60–70 % precision. Food classification of chewing sequences had an average accuracy of 94 %. Mean weight prediction error was lowest for apples (19·4 %) and the largest for lettuce (31 %) |
| Arif | Self-monitoring (physical activity) | Gesture recognition | Recognise physical activity type | Rotation forest classifier | Inertial measurement units (IMUs) placed at wrist, chest and ankles | Accurately (98 %) identified seventeen physical activities ranging from ambulation to activities of daily living |
| Aswani | Self-monitoring (eating behaviour; physical activity), optimise goal setting | Predictive analytics | To predict weight loss based on subject characteristics, step count, energy intake and counselling sessions | Bayesian classification | Secondary data | Predictive modelling framework was competitive in terms of prediction accuracy with linear SVM, logistic regression and decision tree models, which further justified the use of the utility-maximising framework and its ability to capture ‘irrational’ discounting in the decision-making of individuals participating in the intervention |
| Aziz | Self-monitoring (physical activity and energy expenditure) | Gesture recognition | To estimate energy expenditure during sitting, standing and treadmill walking using a smartwatch | LG Urbane Android smartwatch: tri-axial accelerometer, gyroscope and magnetometer | Physical movement | The activity-based models provided 7 % better energy expenditure estimation than the traditional acceleration-based models |
| Bastian | Self-monitoring (physical activity) | Gesture recognition | To discriminate between eight activity classes (lying, slouching, sitting, standing, walking, running and cycling) in a laboratory condition and walking the streets, running, cycling and taking the bus in free-living conditions | Hip-worn triaxial accelerometer | Physical movement | The performances of the laboratory-calibrated algorithm decreased for several activities when applied to free-living data. Recalibrating the algorithm with data closer to real-life conditions improved the detection of overall sitting (sensitivity: laboratory model 24·9 %; recalibrated model 95·7 %). |
| Bi | Self-monitoring (eating behaviour) | Sound recognition | To monitor | High-fidelity microphone is worn on the subject’s neck near the jaw | Acoustic signals | The accuracy of food-type recognition by AutoDietary is 84·9 %, and those to classify liquid and solid food intakes are up to 97·6 and 99·7 %, respectively. |
| Bouarfa | Self-monitoring (physical activity, energy expenditure) | Gesture recognition | Energy expenditure estimation under free-living conditions | A single ear-worn activity recognition (eAR) sensor (built-in triaxial accelerometer) | Physical movement | In free-living settings, ten different types of physical activities (i.e., lying down, standing, computer work, vacuuming, going up and downstairs, slow walking, brisk walking, slow running, fast running and cycling) were predicted |
| Chung | Self-monitoring (eating behaviour and physical activity) | Gesture recognition | To detect the patterns of temporalis muscle activities during food intake and other physical activities | A glasses-type device with an in-built EMG or piezoelectric strain sensor and attaching them directly onto human skin | Skin movement | The average F1 score of the classification among the featured activities was 91·4 %. |
| Dijkhuis | Self-monitoring (physical activity), optimise goal setting | Predictive analytics | To predict the likelihood of a user to achieve a daily personal step goal | Wrist-worn activity tracker, the Fitbit Flex | Physical activity | In 80 % of the individual cases, the random forest algorithm was the best performing algorithm |
| Dobbins | Self-monitoring (physical activity) | Gesture recognition | To distinguishing physical activity | Tri-axial accelerometers and a heart-rate monitor | Physical movement | The results showed an improvement in recognition accuracy as compared with existing studies, with accuracies of up to 99 % and sensitivities of 100 % |
| Dong | Self-monitoring (eating behaviour) | Gesture recognition | To automatically detect periods of eating (free-living condition) | iPhone 4 (accelerometer and gyroscope) placed inside a pouch, wrapped snugly around the forearm | Wrist motion (linear and rotational) | Results show an accuracy of 81 % for detecting eating at 1 s resolution |
| Ermes | Self-monitoring (physical activity) | Gesture recognition | To recognise sports performed by the subjects | 3-D accelerometers on hip and wrist and GPS information | Physical movement | The total accuracy of the activity recognition using both supervised and unsupervised data was 89 % that was only 1 % unit lower than the accuracy of activity recognition using only supervised data. The accuracy decreased by 17 % unit when only supervised data were used for training and only unsupervised data for validation. |
| Everett | Self-monitoring (physical activity), self-control (physical activity), optimise goal setting | Real-time analytics with personalised micro-interventions | Automatically translate various raw mobile phone data into insights about user’s life habits. Provide personalised, contextual, just-in-time, just-in place recommendations (tailored messages) | Phone accelerometer | Physical movement | Physical activity increased by 2·8 metabolic equivalents of task (MET) – hours per week, weight reduced by 1·6 kg ( |
| Fontana | Self-monitoring (eating behaviour) | Gesture recognition | To objectively monitor ingestive behaviour in free-living | A jaw motion sensor, a hand gesture sensor and an accelerometer (integrated into a device and wirelessly interface to a smartphone) | Jaw and hand motion | The system was able to detect food intake with an average accuracy of 89·8 % |
| Forman | Self-monitoring (eating behaviour), self-control (eating behaviour), optimise goal setting | Real-time analytics with personalised micro-interventions | Predict dietary lapses and deliver a targeted intervention designed to prevent the lapse from occurring | Ensemble methods (combining weighted vote of predictions from random forest, Logit. Boost, Bagging, Random Subspace, Bayes Net) | Ecological momentary assessment (EMA) 6 times a day + ad hoc entry of lapse event | Of the twenty-one possible triggers, 29 % of intervention triggers were based on time of day, 16·7 % based on low motivation and 10 % based on fatigue. The remaining eighteen triggers were identified as risk factors < 10 % of the time. There was a reduction in unplanned lapses. Participants averaged a 3·13 % weight loss |
| Forman | Self-monitoring (eating behaviour), self-control (eating behaviour), optimise goal setting | Real-time analytics with personalised micro-interventions | Predict dietary lapses and deliver a targeted intervention designed to prevent the lapse from occurring | Ensemble methods (e.g., combining weighted vote of predictions from Random Forest, Logit. Boost, Bagging, Random Subspace, Bayes Net) | EMA 6 times a day + ad hoc entry of lapse event | Weight Watcher (WW) + OnTrack (OT) participants reported an average of 29·72 ( |
| Forman | Self-monitoring (eating behaviour), self-control (eating behaviour), optimise goal setting | Real-time analytics with personalised micro-interventions | Predict dietary lapses and deliver a targeted intervention designed to prevent the lapse from occurring | Reinforcement learning algorithm | Physical activity measured in minutes of moderate-to-vigorous physical activity (MVPA) using Fitbit Flex or a similar type of Fitbit wrist-worn activity tracker | Proposed system achieved weight losses equivalent to existing human coaching programmes (non-optimised (NO) = 4·42 %, individually optimised (IO) = 4·56 %, group-optimised (GO) = 4·39 %) at roughly one-third the cost (1·73 and 1·77 coaching hours/participant for IO and GO, |
| Fullerton | Self-monitoring (physical activity) | Gesture and image recognition | To recognise activity and sub-category activity types through the use of multiple body-worn accelerometers in a free-living environment | Nine body-worn accelerometers for a day of free living | Physical movement | The recognition accuracy of 97·6 %. Controlled and free-living testing provided highly accurate recognition for sub-category activities (> 95·0 %). Decision tree classifiers and maximum features demonstrated to have the lowest computing time |
| Goldstein | Self-monitoring (eating behaviour) | Predictive analytics | Predict dietary lapses | Ensemble methods (combining weighted vote of predictions from random forest, Logit. Boost, Bagging, Random Subspace, Bayes Net) | EMA 6 times a day + ad hoc entry of lapse event | Participants responded to an average of 94·6 % of EMA prompts (range = 85·2–98·9 %) and compliance remained relatively stable throughout the study |
| Goldstein | Self-monitoring (eating behaviour), self-control (eating behaviour) optimise goal setting | Real-time analytics with personalised micro-interventions | To measure dietary lapses and relevant lapse triggers and provide personalised intervention using machine learning | Decision tree | EMA 6 times a day + ad hoc entry of lapse event | Average of 4·36 lapses per week ( |
| Hegde | Self-monitoring (physical activity) | Gesture recognition | To propose an insole-based activity monitor – SmartStep, designed to be socially acceptable and comfortable | Insole-based sensor system: contains a 3D accelerometer, a gyroscope | Physical movement | The overall agreement with ActivPAL was 82·5 % (compared with 97 % for the laboratory study). The SmartStep scored the best on the perceived comfort reported at the end of the study |
| Hezarjaribi | Self-monitoring (energy intake) | Speech recognition | To facilitate nutrition monitoring using speech recognition and text mining | Smartphone microphone | Speech | Speech2Health achieves an accuracy of 92·2 % in computing energy intake |
| Hossain | Self-monitoring (eating behaviour) | Image recognition | To detect and count bites and chews automatically from meal videos | Video recorder | Video images | Mean accuracy of 85·4 % ± 6·3 % concerning manual annotation was obtained for the number of bites and 88·9 % (± 7·4 %) for the number of chews |
| Hua | Self-monitoring (physical activity) | Gesture recognition | To classify nine different upper extremity exercises | Triaxial IMU | Physical movement (kinematics) | Random forest models with flattened kinematic data as a feature had the greatest accuracy (98·6 %). Using the triaxial joint range of motion as the feature set resulted in decreased accuracy (91·9 %) with faster speeds |
| Huang | Self-monitoring (eating behaviour) | Gesture recognition | Recognise eating behaviour and food type | On-board real-time decision algorithm; chewing detection algorithm; decision trees | Electromyography embedded in wearable glasses connected to the smartphone | 96 % accuracy in detecting chewing and classifying five types of food |
| Jain | Self-monitoring (physical activity) | Gesture recognition | To classify activities using built-in sensors of smartphones | The phone kept in the front pocket of the subject’s trousers, built-in accelerometer and gyroscope sensor | Physical movement | Average activity classification accuracy achieved using the proposed method was 97·12 % |
| Jiang | Self-monitoring (energy intake) | Image recognition | To develop a deep model-based food recognition and dietary assessment system to study and analyse food items from daily meal images (e.g., captured by smartphone) | Existing datasets | Images | The system was able to recognise food items accurately with top-1 accuracy of 71·7 % and top-5 accuracy of 93·1 % |
| Juarascio | Self-monitoring (eating behaviour) | Predictive analytics | To detect changes in HRV to in turn detect the risk of experiencing an emotional eating episode in an ecologically valid setting | Empatica E4 wrist-sensor (photoplethysmography: non-invasive the optical measurement that can derive cardiovascular features from light absorption of the skin) and EMA six prompts per day (participants were also instructed to self-report immediately following an emotional eating episode and answer the same questions) | Heart rate variability (HRV) | Support vector machine (SVM) models using frequency-domain features achieved the highest classification accuracy (77·99 %), sensitivity (78·75 %) and specificity (75·00 %), though were less accurate at classifying episodes (accuracy 63·48 %, sensitivity 62·68 % and specificity 70·00 %) and did not meet acceptable classification accuracy |
| Kang | Self-monitoring (physical activity, energy expenditure) | Gesture recognition | To predict the energy expenditure of physical activities | AirBeat system: built-in patch-type sensor module for wireless monitoring of heart rate, exercise index, ECG and a three-axial acceleration motion detector | Physical movement, HR, exercise index, ECG, humidity, and temperature | RMSE of 0·1893 and |
| Kim | Self-monitoring (physical activity) | Gesture and image recognition | To recognise sedentary behaviour | Two accelerometers (waist over the right hip and right thigh) and a wearable camera (around the neck using a lanyard) | Physical movement | ActivPAL showed the most accurate estimate of total sedentary time with MAPE of 4·11 % and percentage of bias of –3·52 % |
| Korpusik | Self-monitoring (energy intake) | Speech recognition | To automatically extract food concepts (nutrients and energetic intake) from a user’s spoken meal description | Amazon Mechanical Turk (AMT) where Turkers were to record ten meal descriptions | Speech | 83 % semantic tagging accuracy |
| Kyritsis | Self-monitoring (eating behaviour) | Gesture recognition | To automatically detect in-meal food intake cycles using the inertial signals (acceleration and orientation velocity) from an off-the-shelf smartwatch | Off-the-shelf smartwatch (acceleration and orientation velocity) | In-meal bite detection | Achieved the highest F1 detection score (0·913 in the leave-one subject-out experiment) as compared with existing algorithms |
| Lin | Self-monitoring (physical activity, energy expenditure) | Gesture recognition | To recognise physical activities and their corresponding energy expenditure | Motion sensors and an ECG sensor | Physical movement and ECG | Recognition accuracies using decision trees in the cross-validations ranged from 95·52 to 97·70 % |
| Lin | Self-monitoring (physical activity, energy expenditure) | Image recognition | To estimate energy expenditure of physical activity in gyms | Kinect for XBOX 360 sensors (depth and motion sensing) | Physical movement (Kinect skeletal data) | The measured and predicted metabolic equivalents of task exhibited a strong positive correlation |
| Liu | Self-monitoring (physical activity, energy expenditure) | Gesture recognition | To recognise the physical activity | Two triaxial accelerometers (at hip and wrist), and one ventilation sensor secured to the abdomen (AB) at the level of the umbilicus | Ventilation (abdomen), motion (hip), motion (wrist) | Correctly recognised the thirteen activity types 88·1 % of the time, which is 12·3 % higher than using a hip accelerometer alone. Also, the method predicted energy expenditure with a root mean square error of 0·42 MET, 22·2 % lower than using a hip accelerometer alone |
| Liu | Self-monitoring (physical activity and energy expenditure), self-control (eating, physical activity), optimise goal setting | Gesture recognition and real-time analytics with personalised micro-interventions (1 subject) | To recognise physical activity and provide health feedback | Android phone with built-in accelerometer and magnetometer | Nine basic daily physical activities: walking, jogging, ascending and descending stairs, bicycling, travelling up in an elevator, travelling down in an elevator, using an escalator and remaining stationary | Achieved an average recognition accuracy of 98·0 % with a minimised energy expenditure |
| Liu | Self-monitoring (eating behaviour) | Image recognition | To recognise food items | Camera on smartphones | Food type | (1) outperformed existing work in terms of food recognition accuracy (top-1: 77·5 %; top-5: 95·2 %); (2) reduced response time that is equivalent to the minimum of the existing approaches and (3) lowered energy consumption which is close to the minimum of the state of the art |
| Lo | Self-monitoring (eating behaviour) | Image recognition | To estimate the portion size of food items consumed | Camera | Portion size (often commonly seen food categories including burger, fried rice, pizza, etc. Each category has twenty food models with different shape geometries and portion size) | Mean accuracy of up to 84·68 % |
| Lopez-Meyer | Self-monitoring (eating behaviour) | Image recognition | To describe the detection of food intake by a support vector machine classifier trained on-time history of chews and swallows | Videotaped by a camcorder to capture subject activity | Chewing and swallowing | The highest accuracy of detecting food intake (94 %) was achieved when both chews and swallows were used as predictors |
| Mo | Self-monitoring (physical activity, energy expenditure) | Gesture recognition | To estimate energy expenditure | Wireless wearable multi-sensor integrated measurement system (WIMS): two triaxial accelerometers, worn at the hip and wrist | Body motion and breathing | Under free-living conditions, WIMS correctly recognised the activity intensity level 86 % of the time |
| Montoye | Self-monitoring (physical activity) | Gesture recognition | To recognise physical activity type | Oxycon Mobile portable metabolic analyser and four accelerometer-based activity monitors | Physical movement | Overall classification accuracy for assessing activity type was 66–81 % for accelerometers mounted on the hip, wrists and thigh, which improved to 73–87 % when combining similar activities into categories. The wrist-mounted accelerometers achieved the highest accuracy for individual activities (80·9–81·1 %) and activity categories (86·6–86·7 %); accuracy was not different between wrists. The hip-mounted accelerometer had the lowest accuracy (66·2 % individual activities, 72·5 % activity categories) |
| Päßler | Self-monitoring (eating behaviour) | Sound recognition | To recognise chewing sounds | Microphones applied to the outer ear canal | Chewing sound | Precision and recall over 80 % were achieved by most of the algorithms |
| Parkka | Self-monitoring (physical activity) | Gesture recognition | To automatically recognise the physical activity | Nokia wireless motion bands 3-D accelerometer | Physical movement | Overall accuracy was 86·6 and 94·0 % after classifier personalisation |
| Pouladzadeh | Self-monitoring (energy intake) | Image recognition | To estimate food energy and nutrition | The built-in camera of smartphones or tablets | Food size, shape, colour and texture. Food portion was estimated based on the area | Accuracies in detecting single, non-mixed and mixed foods were 92·21, 85 and 35–65 %, respectively |
| Pouladzadeh | Self-monitoring (eating behaviour) | Sound recognition | To estimate food energy and nutrition using a cloud-based support vector machine (SVM) method | Built-in camera of smartphones or tablets | Food size, shape, colour and texture. Food portion was estimated based on the area | By using a cloud computing system in the classification phase and updating the database periodically, the accuracy of the recognition step has increased in single food portion, a non-mixed and mixed plate of food compared with LIBSVM |
| Rabbi | Self-monitoring (physical activity), self-control (physical activity), optimise goal setting | Real-time analytics with personalised micro-interventions | To automatically (1) track physical activity, (2) analyse activity and food logs to identify frequent and nonfrequent behaviours and (3) generate personalised suggestions that ask users to either continue, avoid or make small changes | Accelerometer and GPS, smartphone food logging | Four most common | Physical activity increased by 2·8 metabolic equivalents of task (MET) – hours per week ( |
| Rachakonda | Self-monitoring (energy intake) | Image recognition | To automatically detect, classify and quantify the objects from the plate of the user | A camera attached to glasses | Food type, amount, time of eating | The iLog model has produced an overall accuracy of 98 % with an average precision of 85·8 % |
| Sazonov | Self-monitoring (eating behaviour) | Sound recognition | To detect acoustical swallowing | Throat microphone located over laryngopharynx | Swallowing sounds | Average weighted epoch recognition accuracy for intra-visit individual models was 96·8 % which resulted in 84·7 % average weighted accuracy in detection of swallowing events |
| Sazonov | Self-monitoring (eating behaviour) | Gesture recognition | To detect periods of food intake based on chewing | Piezoelectric strain gauge sensor | Jaw movement | Classification accuracy of 80·98 % and a fine time resolution of 30 s |
| Sazonov | Self-monitoring (physical activity, energy expenditure) | Gesture recognition | To describe the use of a shoe-based wearable sensor system (SmartShoe) with a mobile phone for real-time recognition of various postures/physical activities and the resulting EE | Five force-sensitive resistors (integrated into a flexible insole) and an accelerometer | Physical movement | Results showed a classification accuracy virtually identical to SVM (∼95 %) while reducing the running time and the memory requirements by a factor of > 103 |
| Spanakis | Self-monitoring (eating; emotions) | Predictive analytics | Analyse individual states of a person status (emotions, location, activity, etc.) and assess their impact on unhealthy eating | Classification decision trees; hierarchical agglomerative clustering | EMA 10 times a day + ad hoc entry of lapse event | Participants were clustered into six groups based on their eating behaviour and specific rules that discriminate which conditions lead to healthy |
| Spanakis | Self-monitoring (eating; emotions), self-control (eating) | Real-time analytics with personalised micro-interventions | Analyse user-specific data, highlight most discriminating patterns that lead to unhealthy eating behaviour and providing feedback (personalised warning messages before a possible unhealthy eating event) | Classification decision trees and hierarchical agglomerative clustering | EMA 6 times a day + ad hoc entry of lapse event | Participants reported on average 3·6 eating events ( |
| Stein | Self-monitoring (eating; physical activity; emotions), self-control (eating; physical activity), optimise goal setting | Real-time analytics with personalised micro-interventions | Predict dietary lapses and provide adaptive semi-individualised feedback to users regarding their eating behaviour | Used a previously used algorithm which used decision tree | Chatbot | Percentage of healthy meals increased by 31 % of total meals logged at baseline to 67 % within 21 weeks; the percentage of unhealthy meals decreased by 54 %. Users averaged 2·4 kg or 2·4 % weight loss, and 75·7 % (53/70) of users lost weight in the programme |
| Tao | Self-monitoring (physical activity, energy expenditure) | Gesture and image recognition | To estimate energetic expenditure | Camera, two wearable accelerometers | Physical movement, HR, exercise index, humidity and temperature | The fusion of visual and inertial data reduces the estimation error by 8 and 18 % compared with the use of visual-only and inertial sensor only, respectively, and by 33 % compared with a MET-based approach |
| Thomaz | Self-monitoring (eating) | Sound recognition | Recognise eating behaviour (chewing and biting sound from ambient noises) | SVM; nearest neighbours and random forest | Wrist-worn audio recording device | Detected eating with 86·6 % accuracy |
| Vathsangam | Self-monitoring (physical activity, energy expenditure) | Gesture recognition | To estimate energy expenditure during treadmill walking | Inertial measurement unit (IMU): triple-axis accelerometer, triaxial gyroscopes | Physical movement | Combining accelerometer and gyroscope information leads to improved accuracy compared with using either sensor alone |
| Vathsangam | Self-monitoring (physical activity, energy expenditure) | Gesture recognition | To detect physical activity using different features | Phone-based triaxial accelerometer | Physical movement | Feature combinations corresponding to sedentary energy expenditure, sedentary heart rate and sex alone resulted in errors that were higher than speed-based models and nearest-neighbour models. Size-based features such as BMI, weight and height produced lower errors. Weight was the best individual descriptor followed by height. |
| Walker | Self-monitoring (eating behaviour) | Sound recognition | To automatically detect ingestion | Throat microphone located over laryngopharynx | Eating sound | > 94 % of ingestion sounds are correctly identified with false-positive rates around 9 % based on 10-fold cross-validation |
| Wang | Self-monitoring of weight loss progress | Predictive analytics | Predict weight loss based on socio temporal context | Linear regression; stochastic gradient descent (SGD) | Secondary data | Weight loss can be predicted based on temporal-social information |
| Yunus | Self-monitoring (energy intake) | Image recognition | To automatically estimate food attributes such as ingredients and nutritional value | Existing image datasets | Food type and portion | Results showed the top 1 classification rate of up to 85 % |
| Zhang | Self-monitoring (eating behaviour) | Gesture and image recognition | To detect eating | Wrist-worn sensor (Microsoft Band 2-accelerometer and gyroscope) and an HD webcam camera | Eating and non-eating gestures | Results showed a correlation between feeding gesture count and energetic intake in unstructured eating ( |
| Zhang | Self-monitoring (physical activity) | Wi-Fi signal recognition | To recognise general physical activity | The software platform, a signal transmitter and a signal receiver | Wi-Fi signal | Results showed a recognition rate of the general presence of physical activity of 99·05 %, an average recognition rate of 92 % when detecting four common classes of activities |
| Zhou | Self-monitoring (physical activity) | Predictive analytics | Predict exercise lapse | SVM; logistic regression | Secondary data | Discontinuation prediction score (DiPS) makes accurate predictions on exercise goal lapse based on short-term data. The most predictive features were steps and physical activity intensity |
| Zhou | Self-monitoring (physical activity), self-control (physical activity), optimise goal setting | Real-time analytics with recommendations | Adaptively compute personalised step goals that are predicted to maximise future physical activity for each participant based on all the past steps’ data and goals of each participant | Behavioural analytics algorithm (BAA) | Phone accelerometer | Participants in the intervention group had a decrease in mean ( |
Summary of AI features (that uses machine learning), instruments/sensors, sensing domains and functions about weight management
| AI features | Instruments/sensors | Sensing domains |
|---|---|---|
| Gesture recognition | Accelerometer( | Inertia: arm, hand-to-mouth, jaw movement (bite detection), physical activity |
| Five force-sensitive resistors( | Pressure | |
| Heart rate monitor (patch-type sensor or built-in smartphone/smartwatch)( | Physiological: cardiovascular parameters mainly heart rate, ventilation | |
| Piezoelectric sensor( | Swallowing movement (jaw and skin), chewing cycle and food type | |
| Global positioning system (GPS)( | Location | |
| Image recognition | Video recorder( | Images of food item/group/type |
| Camera( | Images of food size, shape, colour, portion and texture | |
| Sound recognition | Ear microphone (worn on one’s inner ear, outer ear, wrist, neck near the jaw/throat)( | Chewing sound (bone-conducted food breakdown sounds) |
| Speech recognition | Microphone (smartphone in-build microphone)( | Verbal food description or nutrition label |
| Wireless signal recognition | Signal receiver( | Wireless signals |
Details of studies that used real-time analytics with personalised micro-interventions (n 10)
| Author, year | Intervention | Intervention duration; run-in period | Subjective data | Objective data | Feasibility and acceptability | Accuracy |
|---|---|---|---|---|---|---|
| Everett | Sweetech app – uses machine learning to automatically translate raw data streams originating from the patient’s mobile phone and into insights about the individual’s life habit and provides personalised, contextual, just-in-time, just-in-place, recommendations | 12 weeks, NR | Demographic information, past medical history and medications. 4-item Physical Activity States of | Weight: digital body weighing scale (Bluetooth) | 86 % retention; | NR |
| Forman | OnTrack + dietary weight loss programme called Weight Watcher (WW) – uses machine learning algorithm to automatically build models of lapse behaviour, predict lapses before they occur and delivers micro-interventions (messages) when lapse risk is high | 8 weeks; | Behavioural risk factors and lapse behaviour (21: affect, boredom, hunger, cravings, tiredness, unhealthy food availability, temptations, missed meals/snacks, self-efficacy (confidence), motivation, socialising (with or without food present), TV, negative interpersonal interactions, healthy food presence, cognitive load, food cues (advertisements), hours of sleep, exercise, alcohol consumption, planning food intake, time of the day | Height and weights: calibrated scale and stadiometer | 85·1 % completed; 70·15 % opened risk alerts; Technology Acceptance Model Scales (TAMS): M = 6·14, | 72 % accuracy, 70 % sensitivity and 72 % specificity, 80 % negative predictive value |
| Forman | OnTrack + WW | 10 weeks; 2 weeks | Behavioural risk factors and lapse behaviour (seventeen potential lapse triggers, i.e., affect, boredom, hunger, cravings, tiredness, unhealthy food availability, temptations, missed meals/snacks, self-efficacy, socialising, watching TV, negative interpersonal interactions, cognitive load, food cues/advertisements, hours of sleep, alcohol consumption, and planning food intake. Time of day, automatically measured, served as an 18th trigger.) | Weight: Yumani Smart Scale (Bluetooth) | 64·4 % completed; 46·9 % opened risk alerts; TAMS: M = 4·70, | 69·2 % sensitivity; 83·8 % specificity |
| Forman | AI-optimised interventions include individually optimised (i.e., at each of the 24 intervention points, participants receive the intervention with the highest reward score for them so far, except when the system is ‘exploring’) or group-optimised (i.e., interventions are assigned based on the highest possible total reward scores, across all interventions assigned, given a predetermined amount of total intervention time across all participants for the day) | 16 weeks; NR | Energy intake: participants logged all food and beverages using the Fitbit mobile phone application | Weigh: Yumani Smart Scale (Bluetooth) | A short survey of coaches: the portal was easy to use (M = 3·33 out of 4) and able to effectively carry out the remote coaching (M = 3·33 out of 4); | NR |
| Liu | SmartCare – an energy-efficient long-term physical activity tracking system that follows users’ physical activity habits and gives personalised quantitative health assessment and health regime suggestion | 4 weeks; NR | Users’ daily physical activities and body type: nine basic daily physical activities: walking, jogging, ascending and descending stairs, bicycling, travelling up in an elevator, travelling down in an elevator, using | Smartphone built-in accelerometer and magnetometer | NR | 98 % accuracy in physical activity recognition |
| Rabbi | MyBehaviour – (1) uses a combination of automatic and manual logging to track physical activity (e.g., walking, running, gym), user location, and food, (2) automatically analyse activity and food logs to identify frequent and nonfrequent behaviours and (3) generate personalised suggestions that ask users to either continue, avoid or make small changes to existing behaviours to help users reach behavioural goals | 3 weeks, NR | Activity tracking and manual food logging either by selecting food items from a database or directly input energy information from nutrition labels. (Users can take photos of food as reminders to input energy intake) | Accelerometer and GPS | According to the suggestion-rating survey, participants in the experimental group had a significantly higher intention to follow personalised suggestions than those in the control group in following generic suggestions | NR |
| Goldstein | Two OnTrack versions – OnTrack-short (OT-S) (8 lapse trigger questions at each EMA survey) and OnTrack-long (OT-L) (17 lapse triggers questions at each EMA survey). When an EMA survey was completed, the algorithm classified responses as no risk (when a prediction was ‘no lapse’), low risk (probability of lapse > 40 %), medium risk (probability of lapse between 40 and 70 %) or high risk (probability of lapse > 70 %) | 10 weeks; 2 weeks | Seventeen lapse triggers: affect, sleep, fatigue, hunger, motivation to adhere to a diet, cravings, boredom, temptation, cognitive load, confidence, socialising, television, negative interpersonal interactions, presence of tempting foods, food advertisements, planning food, alcohol, time/ | NR | 84·3 % completed; | 79·8% accuracy of lapse prediction (79·7% in OT-S |
| Spanakis | Think Slim – uses machine learning to predict unhealthy eating behaviour and allow users to report potential unhealthy eating promoting factors (emotions, activities, etc.). Emphasis is given to providing feedback before possible unhealthy eating events (i.e., warn users in the appropriate time manner using a classification algorithm) and to construct groups of eating behaviour profiles (using a clustering algorithm) | 8 weeks; 1 week | Fifteen lapse triggers: date, food craving, seven emotions each measured on ten-point VAS scale (worried, angry/annoyed, stressed/tense/relaxed/at ease. Cheerful/happy, sad/depressed, bored), specific craving, location, activity, specific eating, thoughts regarding eating, food intake image/ | NR | 70·5 % completed | NR |
| Stein | Lark’s AI health – uses machine learning to power a Chatbot that mimic health professionals’ empathetic health counselling | 16 weeks; NR | Weight loss, meal quality, physical activity and sleep data were collected through user input | Sleep and physical activity, partly through automatic detection by the user’s mobile phone. User engagement was assessed by duration and amount of app use | 44·0 % active users by end of the intervention; | NR |
| Zhou | CalFit app – mobile phone app which delivers daily step goals using push notifications and allows real-time physical activity monitoring | 10 weeks; 1 week | Socio-demographic information, self-reported medical history, Barriers to Being Active Quiz (twenty-one questions on a ten-point Likert scale on seven sub-areas: lack of time, social influence, lack of energy, lack of willpower, fear of injury, lack of skill, and lack of resources), International Physical Activity Questionnaire – Short Form/ | Phone accelerometer | 77·5 % retention | NR |
NR, not reported; IMU, inertial measurement unit.
Included within intervention;
Fig. 4Proposed mechanism of AI-assisted self-regulation