| Literature DB >> 34946400 |
Ghalib Ahmed Tahir1, Chu Kiong Loo1.
Abstract
Dietary studies showed that dietary problems such as obesity are associated with other chronic diseases, including hypertension, irregular blood sugar levels, and increased risk of heart attacks. The primary cause of these problems is poor lifestyle choices and unhealthy dietary habits, which are manageable using interactive mHealth apps. However, traditional dietary monitoring systems using manual food logging suffer from imprecision, underreporting, time consumption, and low adherence. Recent dietary monitoring systems tackle these challenges by automatic assessment of dietary intake through machine learning methods. This survey discusses the best-performing methodologies that have been developed so far for automatic food recognition and volume estimation. Firstly, the paper presented the rationale of visual-based methods for food recognition. Then, the core of the study is the presentation, discussion, and evaluation of these methods based on popular food image databases. In this context, this study discusses the mobile applications that are implementing these methods for automatic food logging. Our findings indicate that around 66.7% of surveyed studies use visual features from deep neural networks for food recognition. Similarly, all surveyed studies employed a variant of convolutional neural networks (CNN) for ingredient recognition due to recent research interest. Finally, this survey ends with a discussion of potential applications of food image analysis, existing research gaps, and open issues of this research area. Learning from unlabeled image datasets in an unsupervised manner, catastrophic forgetting during continual learning, and improving model transparency using explainable AI are potential areas of interest for future studies.Entities:
Keywords: automatic diet monitoring; feature extraction; food datasets; food recognition; image analysis; interactive segmentation; volume estimation
Year: 2021 PMID: 34946400 PMCID: PMC8700885 DOI: 10.3390/healthcare9121676
Source DB: PubMed Journal: Healthcare (Basel) ISSN: 2227-9032
Figure 1Scope and taxonomy of this survey paper.
Figure 2Confusion matrix.
Food image datasets.
| Authors | Year | Dataset | Food Category | Total # Images/Class | Image Source |
|---|---|---|---|---|---|
| S. Godwin et al. [ | 2006 | Wedge Shape foods dataset | American Foods | 3 categories | Controlled environment |
| Chen et al. [ | 2009 | PFID | American Fast Foods | 1038(61) | Fast food data captured in |
| Mariappan et al. [ | 2009 | TADA | Artificial And | 256(11) | Controlled environment |
| Yanai et al. [ | 2010 | Food-50 | Japanese Foods | 5000(50) | Crawled from web |
| Hoashi et al. [ | 2010 | Food-85 | Japanese Foods | 8500(85) | Existing food databases |
| Miyazaki et al. [ | 2011 | Foodlog | Japanese Foods | 6512(2000) | Captured by users |
| Marc Bosch et al. [ | 2011 | FNDDS | American Foods | 7000 | Images of food accquired by users |
| Matsuda et al. [ | 2012 | UECFOOD-100 | Japanese Foods | 14,361(100) | Captured by mobile camera |
| Chen et al. [ | 2012 | ChineseFoodNet | Chinese dishes. | 192,000(208) | Gathered from web |
| M.-Y. Chen et al. [ | 2012 | Chen | Chinese Foods | 5000/50 | Crawled from the Internet |
| Bossard et al. [ | 2014 | Food-101 | American Foods | 101,000(101) | Crawled from web |
| L. Bossard et al. [ | 2014 | ETHZ Food-101 | American Foods | 100,000(101) | Crawled from web |
| Kawano et al. [ | 2014 | UECFOOD-256 | Japanese Foods | 25,088(256) | Captured by mobile camera |
| T. Stutz et al. [ | 2014 | Rice dataset | Generic (Rice) | 1 food type | Acquired from user |
| Farinella et al. [ | 2014 | UNCIT-FD889 | Italian Foods | 3583 (899) | Acquired with a smartphone |
| Meyers et al. [ | 2015 | FOOD201-Segmented | American Foods | 12625 | Manually annotated dataset |
| Xin Wang et al. [ | 2015 | UPMC Food-101 | Generic | 100,000(101) | Crawled from web |
| Cioccoa et al. [ | 2015 | UNIMB 2015 | Generic | 2000(15) | Using a Samsung Galaxy |
| Shaobo Fang et al. [ | 2015 | TADA(19 foods) | American Foods | 19 categories | Controlled environment |
| Xu et al. [ | 2015 | Dishes | Chinese Restaurant Foods | 117,504(3832) | Download from dianping |
| Beijbom et al. [ | 2015 | Menu-Match | Generic Restaurant Food | 646(41) | Captured from social media |
| Zhou et al. [ | 2016 | Food-975 | Chinese Foods | 37,785(975) | Collected from restaurants |
| J. chen et al. [ | 2016 | Vireo-Food 172 | Chinese Foods | 110,241(172) | Downloaded from web |
| Cioccoa et al. [ | 2016 | UNIMB 2016 | Italian Foods | 1027(73) | Captured from |
| Hui Wu et al. [ | 2016 | Food500 | Generic | 148,408(508) | Crawled from web |
| Singla et al. [ | 2016 | Food-11 | Generic | 16,643(11) | Other food datasets |
| Farinella et al. [ | 2016 | UNCIT-FD1200 | Generic | 4754(1200) | Acquired using smartphone |
| Jaclyn Rich et al. [ | 2016 | Instagram 800k | Generic | 808,964(43) | Social Media |
| Liang et al. [ | 2017 | ECUSTFD | Generic | 2978(19) | Acquired using smartphone |
| Güngör et al. [ | 2017 | Turkish-Foods-15 | Turkish Dishes | 7500/15 | Collected from other datasets |
| Pandey et al. [ | 2017 | Indian Food | Indian Foods | 5000(50) | Downloaded from web |
| Termritthikun et al. [ | 2017 | THFood-50 | Thai Foods | 700/50 | Downloaded from web |
| Ciocca et al. [ | 2017 | FOOD524DB | Generic | 247,636(524) | Existing food database |
| Hou et al. [ | 2017 | VegFru | Generic (Fruit and VEG) | 160,731(292) | Collected from search engine |
| Waltner et al. [ | 2017 | FruitVeg-81 | Generic (Fruit and VEG) | 15,630(81) | Collected using mobile phone |
| Muresan et al. [ | 2018 | Generic (Fruits 360 Dataset) | Fruit Dataset | 71,125(103) | Camera |
| Qing Yu et al. [ | 2018 | FLD-469 | Japanese Foods | 209,700(469) | Smart Phone camera |
| Kaur et al. [ | 2019 | FoodX-251 | Generic | 158,000(251) | Collected from web |
| Ghalib et al. [ | 2020 | Pakistani Food Dataset | Pakistani Dishes | 4928(100) | Crawled from web |
| Narayanan et al. [ | AI-Crowd | Swiss Foods | 25,389 | Volunteer Users | |
| Bolaños M. et al. [ | 2016 | EgocentricFood | Generic | 5038(9) | Taken by a wearable |
| E. Aguilar et al. [ | 2019 | MAFood-121 | Spanish Foods | 21,175 | Google search engine |
Figure 3System Flow.
Figure 4Sample images from few food datasets.
Figure 5Handcrafted feature extraction methods.
Handcrafted features.
| Reference | Year | Visual Features | Dataset | Recognition Type |
|---|---|---|---|---|
| Hoashi et al. [ | 2010 | Bag-of-features (BoF), | Used for recognition | Automatic |
| Yang et al. [ | 2010 | Deals with | Pittsburgh | Food recognition |
| kong and Tan [ | 2011 | SIFT, | Pittsburgh Food Image Dataset (PFID) | Regular shaped |
| Bosh et al. [ | 2011 | Global feature classes: texture and color | Database consisting of | Food recognition |
| Zhang et al. [ | 2011 | Color, SIFT, Shape, RGB histograms | Dataset came from online sources, | Classification |
| Matsuda et al. [ | 2012 | Gabor texture features, | Food image dataset | Multiple |
| Kawano and Yanai [ | 2013 | Bag-of-features and color histogram, HOG | - | Mobile food |
| Anthimopoulos et al. [ | 2014 | Bag-of-features, | Visual dataset consisting of 5000 | Food recognition |
| Tammachat and | 2014 | Bag-of-features (BoF), Texture | Database consisting of 40 types of | Food image |
| Pouladzadeh et al. [ | 2014 | Graph cut, Color and Texture | Dataset consisting of 15 different | Food image |
| He et al. [ | 2014 | Color, Texture, Dominant Color Descriptor (DCD), | Food image dataset containing | Food image |
| Kawano and Yanai [ | 2014 | Color, HoG and Fisher Vector | UECFOOD-256 food image dataset | Real-time food |
| Oliveira et al. [ | 2014 | Color, Texture | Images were gathered using | Mobile Food |
| Pouladzadeh et al. [ | 2015 | Color, Texture, Size, Shape, Gabor filter | System was tested on single food | Cloud-based |
| Farinella et al. [ | 2016 | SIFT, Bag of Textons, PRICoLBP | UNICT-FD1200 dataset. | Food image recognition |
Deep visual features.
| Reference | Year | Features | Dataset | Recognition Type |
|---|---|---|---|---|
| Kawano and Yanai, [ | 2014 | Fisher Vector and DCNN | UECFOOD-100 and | Food image recognition |
| Yanai and Kawano, [ | 2015 | DCNN | UECFOOD-100 | Food image recognition |
| Christodoulidis et al. [ | 2015 | CNN | Manually annotated dataset | Food recognition |
| Pouladzadeh et al. [ | 2016 | Graphcut and DCNN | Database consisting of | Food recognition for |
| Hassannejad et al. [ | 2016 | Inception | Food-101, UECFOOD-100 | Food image recognition |
| Liu et al. [ | 2016 | DCNN | Food-101, UECFOOD-256 | Mobile food image recognition |
| Chen and Ngo, [ | 2016 | Arch-D | Chinese Foods | Ingredient recognition |
| Ciocca et al. [ | 2017 | VGG | UNIMIB 2016 | Food recognition |
| Termritthikun et al. [ | 2017 | NU-InNet | THFOOD-50 | Food recognition |
| Pandey et al. [ | 2017 | AlexNet, GoogLeNet | ETH Food-101 and | Food Recognition |
| Liu et al. [ | 2018 | GoogleNet | UECFOOD-100, UECFOOD-256 | Food recognition |
| McAllister et al. [ | 2018 | ResNet-152, GoogLeNet | Food 5k, Food-11, RawFooT-DB | Food recognition |
| Martinel et al. [ | 2018 | WISeR | UECFOOD-100, UECFOOD-256 | Food recognition |
| E. Aguilar et al. [ | 2018 | AlexNet | UNIMIB2016 | Automatic food tray analysis |
| S. Horiguchi et al. [ | 2018 | GoogleNet | Built their own food dataset FoodLog | Food image recognition |
| Gianluigi Ciocca et al. [ | 2018 | ResNet50 | Food 475 | Food image recognition |
| B. Mandal et al. [ | 2019 | SSGAN | ETH Food-101 and Indian Food Dataset | Food Recognition |
| G.Ciocca et al. [ | 2020 | GoogleNet, Inception-v3, | Own dataset containing 20 different | Food category recognition, |
| L. Jiang et al. [ | 2020 | VGGNet | UECFOOD-100, UECFOOD-256 and | Food recognition and dietary assesment |
| C. Liu et al. [ | 2020 | VGGNet, ResNet | Vireo-Food 172 | Food ingredient recognition |
| H. Liang et al. [ | 2020 | ChineseFoodNet and Vireo-Food 172 | Chinese food recognition | |
| H. Zhao et al. [ | 2020 | VGGNet, ResNet and DenseNet | UECFOOD-256 and Food-101 | Mobile food recognition |
| G. A. Tahir and C. K. Loo [ | 2020 | ResNet-50, DenseNet201 | Pakistani Food Dataset, UECFOOD-100, | Food recognition |
| C. S. Won [ | 2020 | ResNet50 | UECFOOD-256, Food-101 | Fine grained |
| Zhidong Shen et al. [ | 2020 | Inception-v3, Inception-v4 | Dataset was created including | Food recognition |
Traditional machine learning methods for food category classification.
| Reference | Year | Classification Technique | Classification Accuracy | |
|---|---|---|---|---|
| Top 1 | Top 5 | |||
| Hoashi et al. [ | 2010 | Multiple Kernel Learning (MKL) | Own Food Dataset = 62.5% | N/A |
| Yang et al. [ | 2010 | Support Vector Machine (SVM) | PFID = 78.0% | N/A |
| Kong and Tan [ | 2011 | Multi-class SVM | PFID = 84% | N/A |
| Bosh et al. [ | 2011 | Support Vector Machine (SVM) | Dataset collected = 86.1% | N/A |
| Zhang et al. [ | 2011 | SVM regression with RBF kernel | Own Food Dataset = 82.9% | N/A |
| Matsuda et al. [ | 2012 | Multiple Kernel Learning (MKL) | Own food Dataset = 55.8% | N/A |
| Kawano and Yanai [ | 2013 | Linear SVM and fast tookernel | N/A | 81.6% |
| Anthimopoulos et al. [ | 2014 | Linear SVM | Own Food Dataset = 78.0% | N/A |
| Tammachat and | 2014 | Support Vector Machine (SVM) | Own Food Dataset = 70.0% | N/A |
| Pouladzadeh et al. [ | 2014 | Support Vector Machine (SVM) | Own Food Dataset = 95% | N/A |
| He et al. [ | 2014 | K-nearest Neighbors | Own Food Dataset = 64.5% | N/A |
| Kawano and Yanai [ | 2014 | One-vs-rest | UECFOOD-256 = 50.1% | UECFOOD-256 = 74.4% |
| Oliveira et al. [ | 2014 | Support Vector Machine (SVM) | Own Food Dataset | N/A |
| Pouladzadeh et al. [ | 2015 | Cloud-based Support Vector Machine | Own Food Dataset = 94.5% | N/A |
| Farinella et al. [ | 2016 | Support Vector Machine (SVM) | UNICT-FD1200 = 75.74% | UNICT-FD1200 = 85.68% |
Deep learning models for food category classification.
| Reference | Year | Classification Technique | Classification Performance | |
|---|---|---|---|---|
| Top 1 | Top 5 | |||
| Yanai and Kawano [ | 2015 | DCNN | UECFOOD-100 = 78.8% | N/A |
| Christodoulidis et al. [ | 2015 | DCNN | Own dataset = 84.9% | N/A |
| Chen and Ngo [ | 2016 | DCNN | ||
| Pouladzadeh et al. [ | 2016 | DCNN + Graph cut | Own dataset = 99% | N/A |
| Hassannejad et al. [ | 2016 | DCNN | ETH Food-101 = 88.3% | ETH Food-101 = 96.9% |
| Liu et al. [ | 2016 | CNN | UECFOOD-100 = 76.3% | UECFOOD-100 = 94.6% |
| Pandey et al. [ | 2017 | Ensemble Net | ETH-Food101 = 72.1% | ETH-Food101 = 91.6% |
| Ciocca et al. [ | 2017 | CNN | UNIMIB 2016 = 78.3% | N/A |
| Termritthikun et al. [ | 2017 | CNN | THFOOD-50 = 69.8% | THFOOD-50 = 92.3% |
| McAllister et al. [ | 2018 | CNN+ANN+SVM+ | Food-5K = 99.4% | N/A |
| Liu et al. [ | 2018 | DCNN | UECFOOD-256 = 54.5% | UECFOOD-256 = 81.8% |
| Martinel et al. [ | 2018 | DNN | UECFOOD-100 = 89.6% | UECFOOD-100 = 99.2% |
| E. Aguilar et al. [ | 2018 | CNN+SVM | UNIMIB 2016 = 90.0% | N/A |
| Gianluigi Ciocca et al. [ | 2018 | CNN | Food-475 = 81.6% | Food-475 = 95.5% |
| S. Horiguchi et al. [ | 2018 | Sequential Personalized Classifier | FoodLog = 40.2% | FoodLog = 56.6% |
| B. Mandal et al. [ | 2019 | Generative Adversarial Network | ETH Food-101 = 75.3% | ETH Food-101 = 93.3% |
| Aguilar-Torres et al. [ | 2019 | CNN based on ResNet-50 | MAFood-121 = 81.62% | N/A |
| Kaiz Merchant and Yash Pande [ | 2019 | Inception V3 | ETHZ Food-101 = 70.0% | N/A |
| Mezgec, S. et al. [ | 2019 | Deep Learning | Own Food dataset = 93% | N/A |
| L. Jiang et al. [ | 2020 | DCNN (Faster R-CNN) | FOOD20-with-bbx = 71.7% | FOOD20-with-bbx = 93.1% |
| C. Liu et al., 2020 [ | ||||
| H. Zhao et al. [ | 2020 | JDNet | UECFOOD-256 = 84.0% | UECFOOD-256 = 96.2% |
| G. A. Tahir and C. K. Loo [ | 2020 | Adaptive Reduced Class | Food-101 = 87.3% | N/A |
| C. S. Won [ | 2020 | Three-scale CNN | UECFOOD-256 = 74.1% | UECFOOD-256 = 93.2% |
| Zhidong Shen et al. [ | 2020 | CNN | Own dataset = 85.0% | N/A |
| Jiangpeng He et al. [ | 2020 | 18 layer ResNet | Own dataset = 88.67% | N/A |
| Eduardo Aguilar et al. [ | 2020 | CNN | Own dataset = 88.67% | N/A |
| Dario Ortega Anderez et al. [ | 2020 | CNN | Own dataset = 97.10% | N/A |
| G. Song et al. [ | 2020 | CNN | Web crawled dataset = 56.47% | Web crawled dataset = 60.33 |
| Limei Xiao et al. [ | 2021 | CNN | Own dataset = 97.42% | N/A |
| Lixi Deng et al. [ | 2021 | ResNet-50 | School lunch dataset = 95.3% | N/A |
Proposed methods for food ingredient classification.
| Reference | Year | Dataset | Method | Recall | Precision | F1 |
|---|---|---|---|---|---|---|
| Chen et al. [ | 2016 | Vireo-Food 172 | Arch-D | - | - | 67.17% (Micro-F1) |
| UECFOOD-100 | Arch-D | - | - | 82.06% (Micro-F1) | ||
| Bolaños et al. [ | 2017 | Food-101 | ResNet50+ | 73.45% | 88.11% | 80.11% |
| Recipe 5k | ResNet50+ | 19.57% | 38.93% | 26.05% | ||
| Recipe 5k | Inception-v3+ | 42.77% | 53.43% | 47.51% | ||
| Wang, Yunan, et al. [ | 2019 | Economic Rice | Inception-V4 + NS | 71.90% | 72.10% | 71.40% |
| Economic Behoon | Inception-V4 + NS | 77.60% | 68.50% | 69.70% | ||
| Salvador, Amaia, et al. [ | 2019 | Recipe 1M | CNN | 75.47% | 77.13% | 48.61% |
| J. Chen et al. [ | 2021 | VireoFood-172 | DCNN | - | - | 75.77% (Micro-F1) |
Figure 6Food Volume Estimation Methods.
Comparison of single-view methods for food volume estimation.
| Reference | Year | Dataset | Results (E: Error%) | Technique |
|---|---|---|---|---|
| S. Fang [ | 2015 | 19 food items | E: <6% | 3D parameters and reference objects to compute density for estimating the weight of food item |
| Y. He [ | 2013 | 1453 food images | E: 11% (beverages) 63% | “Integrated image segmentation and identification system” |
| T. Miyazaki [ | 2011 | 6512 images | E: 40% | Linear estimation |
| Beijbom, O [ | 2015 | 646 images, with 1386 tagged food items across 41 categories | E: 232 ± 7.2 | Restaurant-specific food recognition considers meal as a whole entry with all of its nutrients details in DB to solve the volume estimation problem for the restaurant scenario. |
| Koichi Okamoto [ | 2016 | 20 kinds of Japanese Foods (60 test image) | E: 21.30% | Single-image-based food calorie estimation system which uses reference objects to determine food region and quadratic curve estimation from the 2D size of foods to their calories |
| Pettitt, C [ | 2016 | Test data from N:6 participants who completed food diary during pilot sudy by wear micro camera | E: 34% | Wearable micro camera in conjunction with food dairies |
| Akpa Akpro Hippocrate [ | 2016 | 119 food images | E: 6.87% | Image processing with cutlery |
| Jia, W. Y [ | 2012 | 224 pictures | E: <10% | 3D location of a circular feature from a 2D image |
| Yang, Y. Q [ | 2011 | 72 images | E: −3.55% | Single digital image, plate reference |
| Huang, J [ | 2015 | fruits (n:6) | imaging processing | |
| Yue, Y [ | 2012 | 6 food replicas | E: Length (−1.18) | A mathematical model based system involves a camera, circular object in a 3D space to compute food volume. |
| Zhang, W [ | 2015 | 15 different kinds of foods | 85% | Portion estimation by counting pixels |
| Rob Comber [ | 2016 | 6 different meals | “Beef ( | Visual Assessment |
| S. Fang [ | 2016 | 10 objects | “3D geometric models and depth images.” | |
| Godwin, S. [ | 2006 | Five portions of 9-inch cake, Seven portions of pizza, Pies were 9 or 10 inches | E: 25% | Estimated portion sizes using a ruler and the adjustable wedge |
| Hernández, Teresita [ | 2006 | 101 subjects, 5 foods | E: 4.8% ± 1.8% | Digital photographs printed onto a poster. |
| Yang et al. [ | 2021 | Virtual Food Dataset and Real Food Dataset (RFD) (1500 images) | E: <9% on VFD, E: 11.6% and 20.1% on RFD. | Estimates volume by computing inner product between the probability vector from modified MobileNetV2 and the reference volume vector. |
| Graikos et al. [ | 2021 | EPIC-KITCHENS and their own food video datasets | 46.32% average MAPE on 16 test foods and 36.90% average MAPE on 6 combined meals. | Generate 3-dimensional point cloud by using depth map, segmentation mask and camera parameters. It then approximates the volume with points cloud-to-volume algorithm. |
| Lo, F.P.W et al. [ | 2019 | Test dataset: 11 food items | E: 15.32%. | 3D point cloud completion from RGB and depth images. |
Comparison of multi-view methods for food volume estimation.
| Reference | Year | Dataset | Results (E: Error%) | Technique |
|---|---|---|---|---|
| F. Zhu [ | 2010 | 3000 images | E: 1% 19 food items (97.2%) | “Camera calibration step and a 3D volume reconstruction step” |
| Xu Chang [ | 2013 | 14 to 20 images for multi-view method | E: 7.4% to 57.3% | Multi-view volume estimation using “Shape from Silhouettes” to estimate the food portion size |
| Kong, Fanyu [ | 2015 | 6 food items | 84–91% | Multi-View RGB images for 3D reconstruction to estimate the volume |
| Trevno, Roberto [ | 2015 | 120 students (n = 120 meals; 57 breakfast + 63 lunch) | 74% (reliability) | Digital Food Imaging Analysis (DFIA) |
| Jia, W. Y [ | 2014 | 100 food samples | E: −2.80% 30% | ebutton is used for taking pictures, and then portion size is calculated semi-automatically by using computer software |
| Xu, C [ | 2013 | E: 10% | 3D MODELLING AND POSE ESTIMATION | |
| Rhyner, D [ | 2016 | 6 meals | 85.10% | Multi-View RGB images, reference card and 3D model for volume estimation |
| T. Stutz [ | 2014 | Rice, blinded servings | E: <33% | Mobile Augmented Reality System |
| Makhsous et al. [ | 2020 | 8 food items tested | 40% improvement in the accuracy of volume estimation as compared to manual calculation. | Employs a mobile Structured Light System (SLS) to measure the food volume and portion size of a dietary intake. |
| Yuan et al. [ | 2021 | Test dataset: 6 food items | E: 0.83 5.23%. | 3D reconstruction from multi-view RGB images. |
| Lo, F.P.W et al. [ | 2019 | Test dataset: 11 food items | E: 15.32%. | 3D point cloud completion from RGB and depth images. |
Figure 7Strengths and weaknesses of automatic food estimation methods.
Summary of feature extraction and classification methods used by existing mobile applications.
| Reference | Year | Application | Food | Feature Extraction | Classification |
|---|---|---|---|---|---|
| Aizawa et al. [ | 2013 | FoodLog | No | Color, SIFT and | Adaboost |
| Oliveira et al. [ | 2014 | - | Yes | Color and Texture | Support Vector Machine (SVM) |
| Probst et al. [ | 2015 | - | - | SIFT, LBP and Color | Linear SVM |
| Meyers et al. [ | 2015 | Im2Calories | Yes | GoogleNet CNN | GoogleNet CNN model |
| Ravi et al. [ | 2015 | FoodCam | No | HoG, LBP and | Linear SVM |
| Waltner et al. [ | 2017 | - | Yes | RGB, HSV and | Random Forest |
| Mezgec and Seljak [ | 2017 | - | - | NutriNet | NutriNet |
| Pouladzadeh et al. [ | 2017 | - | Yes | CNN | Caffe |
| Waltner et al. [ | 2017 | - | Yes | CNN | CNN |
| Ming et al. [ | 2018 | DietLens | - | ResNet-50 CNN | ResNet-50 CNN |
| Jiang et al. [ | 2018 | - | Yes | Colors, Lines, | Reverse Image Search |
| Jianing Sun et al. [ | 2019 | Food Tracker | Yes | DCNN | DCNN |
| G. A. Tahir | 2020 | MyDietCam | Yes | ResNet-50, | Adaptive Reduced |
Figure 8The application provides the top prediction result. This picture is taken from the study of Ghalib et al., 2020 (permission has been obtained from original author).
Figure 9Percentage of datasets summarized according to the types of food. Generic refers to the multi-cultural dataset.
Figure 10Percentage of studies summarized according to the type of feature extraction methods.
Figure 11Volume estimation methods using single images vs. multiple images.
Figure 12Percentage of studies summarized according to the type of methods employed for feature extraction from food images and the category of classifier used for food image analysis in a mobile application.