Literature DB >> 33095178

Accuracy of Mobile Device-Compatible 3D Scanners for Facial Digitization: Systematic Review and Meta-Analysis.

Hang-Nga Mai1, Du-Hyeong Lee1,2.   

Abstract

BACKGROUND: The accurate assessment and acquisition of facial anatomical information significantly contributes to enhancing the reliability of treatments in dental and medical fields, and has applications in fields such as craniomaxillofacial surgery, orthodontics, prosthodontics, orthopedics, and forensic medicine. Mobile device-compatible 3D facial scanners have been reported to be an effective tool for clinical use, but the accuracy of digital facial impressions obtained with the scanners has not been explored.
OBJECTIVE: We aimed to review comparisons of the accuracy of mobile device-compatible face scanners for facial digitization with that of systems for professional 3D facial scanning.
METHODS: Individual search strategies were employed in PubMed (MEDLINE), Scopus, Science Direct, and Cochrane Library databases to search for articles published up to May 27, 2020. Peer-reviewed journal articles evaluating the accuracy of 3D facial models generated by mobile device-compatible face scanners were included. Cohen d effect size estimates and confidence intervals of standardized mean difference (SMD) data sets were used for meta-analysis.
RESULTS: By automatic database searching, 3942 articles were identified, of which 11 articles were considered eligible for narrative review, with 6 studies included in the meta-analysis. Overall, the accuracy of face models obtained using mobile device-compatible face scanners was significantly lower than that of face models obtained using professional 3D facial scanners (SMD 3.96 mm, 95% CI 2.81-5.10 mm; z=6.78; P<.001). The difference between face scanning when performed on inanimate facial models was significantly higher (SMD 10.53 mm, 95% CI 6.29-14.77 mm) than that when performed on living participants (SMD 2.58 mm, 95% CI 1.70-3.47 mm, P<.001, df=12.94).
CONCLUSIONS: Overall, mobile device-compatible face scanners did not perform as well as professional scanning systems in 3D facial acquisition, but the deviations were within the clinically acceptable range of <1.5 mm. Significant differences between results when 3D facial scans were performed on inanimate facial objects and when performed on the faces of living participants were found; thus, caution should be exercised when interpreting results from studies conducted on inanimate objects. ©Hang-Nga Mai, Du-Hyeong Lee. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 23.10.2020.

Entities:  

Keywords:  accuracy; facial digitization; facial scanners; meta-analysis; systematic review

Mesh:

Year:  2020        PMID: 33095178      PMCID: PMC7647818          DOI: 10.2196/22228

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   5.428


Introduction

Oral and facial rehabilitation involves comprehensive diagnosis and treatment planning [1,2]. Facial morphology assessment is vital for the diagnosis of maxillofacial anomalies, surgery, fabrication of prostheses, and postoperative evaluation [2,3]. Esthetics and prognosis of treatment outcomes can be improved through simulation performed on the 3D facial models of patients [4]. The conventional method for generating facial models of patients is physical facial impression, in which a replica of the face is fabricated using elastomeric materials and a gypsum cast [5,6]. However, the method is uncomfortable for patients because their face is covered with materials during the impression-taking process [6]. In addition, the dimensional accuracy of the physical facial impression model is affected by several factors, including the viscosity of the impression materials, setting time, storage conditions, and time interval from material mixing to stone pouring of the casts [7,8]. Furthermore, the human face is made up of complex anatomical structures with complicated skin textures and colors, which makes realistic replication of the face challenging. Modern digital technologies have revolutionized the facial impression method by enabling 3D facial morphology to be captured using noncontact optical facial scanning devices [9,10]. Digital impression does not require conventional laboratory work or the use of impression materials, thus reducing the discomfort and chair time of the patients. Compared with facial stone casts, wherein only direct anthropometric measurements of the faces can be performed for facial analyses, virtually reconstructed models of the face can be utilized for multidisciplinary purposes [11-13]. Facial landmarks can easily be extracted from a digital facial model, and the digitized data format enables image merging and advanced dimensional analyses, such as surface-to-surface distance measurements and volume misfit evaluations, using analytical computer software [3,14-17]. In addition, digital facial scanning provides an efficient basis for dental education and facial recognition [18-20]. Stationary facial scanning systems based on stereophotogrammetry technology were first introduced in dentistry [21]. However, because of the encumbrance and high cost of this technology, handheld scanning systems using laser or structured-light technology were developed [21-23]. Although most professional handheld scanners are considered acceptable in terms of their scan image quality, they are expensive and often require considerable training time to learn their complex scanning protocols [3,24,25]. Alternatively, 3D sensor cameras based on structured-light technology have been developed for smartphone and tablet devices [15,26-28]. An advantage of using mobile devices for face scanning is their user-friendly operation; this reduces the training time for users [15,29]. Apps can be developed and customized for specific purposes by using open source scripts and software coding [15,29]. Moreover, when an external attachment-type 3D sensor camera is used, the position of the camera is controllable in the mobile-device system [27,29]. Facial scanning using a mobile device 3D sensor camera has been attracting a lot of interest in recent years because it is highly portable and cost-effective and because of the popularity of mobile devices [29]. Smartphone- and tablet-compatible 3D facial scanners have been reported to be an effective tool for clinical use in prosthodontic treatment [27,30-33]. However, the accuracy of the digital facial impression obtained with mobile device–compatible face scanners has not been explored. The purpose of this systematic review and meta-analysis was to investigate the accuracy of mobile device–compatible face scanners for facial digitization.

Methods

Study Design

This study was designed based on PRISMA guidelines (Preferred Reporting Items For Systematic Reviews and Meta-Analyses) [34]. This review was not preregistered on PROSPERO. Accuracy was defined as a dimensional discrepancy between the digital facial impression made by a mobile device–compatible face scanning camera and reference image data set. The PICO (population, intervention, comparison, and outcomes) question was as follows: Are digital facial impressions (population) obtained with mobile device–compatible 3D facial scanning cameras (intervention) equivalent to those of professional handheld face scanners (comparison) in terms of accuracy (outcomes)?

Search Strategy

Peer-reviewed studies published until May 27, 2020 were searched using the following formulated Boolean operator: (digital facial impression OR 3D virtual face OR digital face) AND (optical scanner OR 3D scanner OR stereophotogrammetry OR structured light OR laser scanner OR depth sensor cameras OR depth-sensing cameras) AND (smart device OR mobile OR smartphone OR tablet OR notebook OR laptop) AND (validation OR comparison OR accuracy OR agreement OR reliability OR precision OR reproducibility). The Boolean operator was applied in major electronic databases including PubMed (MEDLINE), Scopus, Science Direct, and Cochrane Library. The Google Scholar search engine was used to find additional articles by combining the related MeSH (Medical Subject Headings) terms and text words. No automatic limiter setting was used during the searches to prevent unwanted filtering of related articles. EndNote software (version 9.2, Clarivate Analytics Inc) was used to manage the articles’ references.

Inclusion and Exclusion Criteria

Inclusion and exclusion criteria were set based on the study design, objectives, interventions, and measurement results. The search was limited to articles published in English only. The inclusion criteria for meta-analysis were low risk of bias, low concern for applicability, and relevant numeric data for pool-weighted estimation using the Cohen d statistical method. Accordingly, randomized and nonrandomized controlled trials, cohort studies, case-control studies, and cross-sectional studies that were performed with human participants and on inanimate objects, reporting quantitative assessments of digital facial models obtained with 3D facial scanners and mobile device–compatible 3D facial scan cameras were included in this review. Conversely, conference papers, case reports, case letters, epidemiologic studies, and author or editorial opinion articles were excluded. Original studies that used only 2D images or did not include mobile device–compatible 3D facial scanners were not reviewed, and studies in which the accuracy could not be quantitatively determined were not considered for analysis.

Data Collection

Two reviewers (H-NM and D-HL) independently participated in collecting, screening, and selecting the potential studies based on the information provided by the titles and abstracts. The full texts of relevant articles were assessed and reviewed by both reviewers. The papers that satisfied all the inclusion criteria were considered eligible for review. The following information was collected from full-text papers and recorded on an electronic spreadsheet (Office Excel, Microsoft Inc): authors, year of publication, study purpose, participant information (sample size, mean age, age range, and gender proportion), scanning methods (scanning device, capture technology, working condition, and scanning process), reference standard for validation (direct anthropometry or another 3D scanning device), types of measurement performed (linear distances or surface-to-surface deviation), number of measurements (number of landmarks, measurement times, and raters), measurement results (mean, estimation errors, and types of statistical analysis), and major conclusions. Articles with missing data or unreliable data were excluded from the meta-analysis. The agreement (κ) between the 2 reviewers was calculated. In case of disagreement, a discussion between the 2 reviewers was conducted to resolve the issues.

Quality Assessment and Meta-Analysis

The risk of bias and concern for applicability based on 4 bias domains—patient selection, index test, reference standard, and flow and timing—were assessed by the 2 reviewers using the Quality Assessment Tool for Diagnostic Accuracy Studies-2 (QUADAS-2) [35]. The random- or fixed-effects model was used to analyze the standardized mean difference (SMD) between the experimental and reference data sets to investigate the effect size estimate and the confidence intervals of SMDs using Cohen d [36]. Heterogeneity was evaluated using the Cochran Q test based on the Higgins I2 statistic [37], where a higher I2 value indicated a stronger heterogeneity. When the Q test indicated high heterogeneity across studies (P<.05) or I2>50%, the random-effects model was selected, and subgroup analysis was performed [38]. The subgroup was defined based on the participants or inanimate objects investigated. Publication bias was assessed using the Egger linear regression statistical test and visually inspected using funnel plots. Meta-analyses were performed using the meta package for R software (version 3.6.0, R Foundation for Statistical Computing Platform); the significance level was set at .05. The robvis package (version 0.3.0) was used to visualize the risk-of-bias assessment results [39].

Results

Search Results

The search resulted in a total of 3942 articles, which were reduced to 3726 articles after removing 216 duplicates. In the title screening process, 3674 articles that were outside the scope of this review were excluded, thereby leaving 52 articles for abstract screening. After the exclusion of 24 articles with irrelevant abstracts, the full texts of 28 articles were read and assessed, and 11 articles were considered eligible for this review. Of these, 6 articles were included in the global meta-analysis, 4 articles were included in the living person face subgroup analysis, and 3 articles in the inanimate face subgroup analysis. The results of the searching and screening process are summarized in Figure 1. There was substantial interrater agreement (κ=0.90).
Figure 1

PRISMA flow diagram summarizing search strategy and search results.

PRISMA flow diagram summarizing search strategy and search results.

Quality Assessment and Applicability Concerns

The quality assessment results from the Quality Assessment Tool for Diagnostic Accuracy Studies-2 showed that among the 11 studies included, one study [40] had a high risk of bias, and another study [41] had a high concern for applicability (Figure 2). There were 2 studies [41,42] showing some risk of bias, and there were 2 studies [40,42] for which there were some concerns for applicability. The patient selection and index test had a higher risk of bias than those of other domains in some studies because of unclear statements regarding the methods employed for random sampling [28,43] or the small number of participants included [5,15]. For applicability, the major concerns arose in the index test domain because several studies did not describe the scanning procedures in detail or did not provide sufficient information about the scanning devices [27,28,40,41].
Figure 2

Quality assessment results according to Quality Assessment Tool for Diagnostic Accuracy Studies-2 guidelines.

Quality assessment results according to Quality Assessment Tool for Diagnostic Accuracy Studies-2 guidelines.

Study Characteristics

Extracted data were organized according to the characteristics of the studies (Table 1). The characteristics of the mobile device–compatible face scanners that were investigated are summarized in Multimedia Appendix 1.
Table 1

Characteristics of the included studies.

StudyParticipant or specimenMobile device, face scannerReferenceLandmark, nMeasurementMajor findings
Amornvit (2019)[28]1 mannequin headiPhone X (Apple Inc), FaceApp (Bellus3D Inc)Manual measurementN/AaΔx, Δy, and ΔzΔx: range 10-50 mm; Δy: range 50-120 mm; failure to record the details in the z-axis
Aswehlee (2018)[5]1 impression castScanify (Fuel 3D Technologies Ltd)CTb3D point cloudsRMSEcThe most accurate noncontact 3D digitizer for maxillofacial defects was Vivid 910 (Minolta Corp), followed by Danae (NEC Engineering), 3dMD (3dMD LLC), and Scanify (P<.001).
Elbashti (2019)[15]1 impression castiPhone 6 (Apple Inc), 123D Catch App (Autodesk Inc)CT3D point cloudsRMSESmartphone 3D modeling was not as accurate as that of the commercially available laser scanning, with higher RMSE values in the defect area representing the depth of the defect.
Knoops (2017)[29]8 (4 male, 4 female)Structure Sensor (Occipital Inc)SPd3D point clouds; 4RMSERMSE of the Structure Sensor was significantly higher than that of M4D Scan (Rodin4D) (P=.008). Structure Sensor lacks hardware and software to accurately characterize areas with complex shape and high curvature but is good at describing general facial forms.
Koban (2019)[40]4 cadaver heads (N/A)Sense (3D Systems Inc); iSense (3D Systems Inc)N/A3D point cloudsRMSEArtec Eva (Artec Group) provided significantly more accurate results than those of the Sense (P<.001) and the iSense devices (P<.001). The Sense was more accurate than the iSense scanner; however, the difference was not significant (P=.12).
Koban (2020)[44]30 (15 male, 15 female), 1 mannequin headSense (3D Systems Inc)SP3D point cloudsRMSEWhole face <1.0 mm (RMSE 0.516, SD 0.109 mm).
Liu (2019)[43]2 impression cast (male)Scanify (Fuel 3D Technologies Ltd)CT1311 linear deviations (Δx, Δy, and Δz)Overall, linear deviations <1 mm for Scanify. The mean overall difference <0.3 mm between Scanify (mean 0.74, SD 0.089 mm) and Vectra (mean 0.15, SD 0.015 mm) images.
Maues (2018)[41]10 (5 male, 5 female)Kinect (Microsoft Inc)SP107 linear distances(mean difference)Mean difference between scanning methods was 0.3 (SD 2.03 mm), showing reasonable accuracy.The mean difference between the images taken with Kinect) was 0.1 (SD 0.6 mm; P<.05) showing good accuracy. Kinect appears to be an interesting and promising resource for facial analysis.
Piedra-Cascón (2020)[27]10 (2 male, 8 female)Face Camera Pro (Bellus3D Inc)Manual measurement6RMSEFace Camera Pro exhibited a trueness RMSE of 0.91 mm and a precision RMSE of 0.32 mm.
Ross (2018)[45]16 (8 male, 8 female)iPhone 7 (Apple Inc), Camera+ app (tap tap tap LLC); RealSense (Intel Corp)Structured light3D point cloudsRMSENo significant differences in RMSE values between iPhone scans with 90 photographs (RMSE 1.4, SD 0.6 mm), 60 photographs (RMSE 1.2, SD 0.2 mm), or 30 photographs (RMSE 1.2, SD 0.3 mm). RealSense had significantly higher RMSE than the iPhone experimental groups (P<.001).
Ten Harkel (2017)[42]34 (10 male, 24 female)RealSense (Intel Corp)SP3D point cloudsRMSERealSense depth accuracy was not affected by facial palsy (RMSE 1.48, SD 0.28 mm) compared to a healthy face (RMSE 1.46, SD 0.26 mm) or Sunnybrook posese (P=.76). However, distance of the patients to the RealSense device was shown to affect accuracy, where the highest depth accuracy (1.07 mm) was measured at a distance of 35 cm.

aN/A: not applicable.

bCT: computed tomography.

cRMSE: root-mean-square error (surface-to-surface).

dSP: stereophotogrammetry.

eSunnybrook poses are a facial grading system for evaluating facial movement outcomes, both at rest and through 5 facial expressions based on voluntary movements (forehead wrinkle, gentle eye closure, open mouth smile, snarl, and lip pucker) [42].

Characteristics of the included studies. aN/A: not applicable. bCT: computed tomography. cRMSE: root-mean-square error (surface-to-surface). dSP: stereophotogrammetry. eSunnybrook poses are a facial grading system for evaluating facial movement outcomes, both at rest and through 5 facial expressions based on voluntary movements (forehead wrinkle, gentle eye closure, open mouth smile, snarl, and lip pucker) [42]. Among the 11 studies included, 6 were conducted on adult volunteers or patients [27,29,41,42,44,45] with a mean age of 35.50 years (SD 8.50; range 24-59). The number of participants in these studies ranged from 8 to 34, with 2 to 15 male and 4 to 15 female participants. The other 5 studies were conducted using inanimate objects such as impression casts of the face [5,15,43] or mannequin heads [28,44], and 1 study [40] was conducted on human cadaver heads. Stereophotogrammetry [29,41,42,44], computed tomography [5,15,43], and high-resolution structured-light handheld scanning [40,45] were used as the reference measurements for comparison, and 2 studies [27,28] used manual interlandmark distance as the reference measurement. For the evaluation, most studies [5,15,27,29,40-42,44,45] measured the global surface-to-surface deviation between the reference and test images by calculating the root-mean-square error (RMSE) of the superimposed 3D images using analytical computer software, with a higher RMSE value indicating a higher surface deviation; however, 3 studies [28,41,43] compared the distances between facial landmarks on a digitized face with those between respective landmarks on a physical model obtained using the manual measurement method. Among them, 1 study [41] evaluated both the global surface-to-surface deviation and interlandmark linear distances, and the deviation was assessed along the x-axis (horizontal length), y-axis (vertical length), and z-axis (depth) in another study [28].

Meta-Analysis

The global analysis revealed heterogeneity (I2=91% P<.001). Random-effects models were selected for both global and subgroup meta-analyses based on the heterogeneity among the studies. In general, the accuracy of facial models obtained with mobile device–compatible face scanners was significantly lower than that of facial models obtained using professional face scanners (SMD 3.96 mm, 95% CI 2.81-5.10 mm, z=6.78, P<.001; Figure 3). Results from the subgroup analysis revealed a significant difference between the subgroups (Figure 4). The difference between the mobile device–compatible and professional face scanners was significantly higher for the face scans of inanimate facial objects (SMD 10.53 mm, 95% CI 6.29-14.77 mm) than for those of living participants (SMD 2.58 mm, 95% CI 1.70-3.47 mm, P<.001, df=12.94).
Figure 3

Global meta-analysis results of comparison of facial models obtained using mobile device–compatible face scanners versus professional face scanners.

Figure 4

Subgroup meta-analysis results of comparison of facial models obtained using mobile device–compatible face scanners versus professional face scanners. (a) 3D facial scans performed on inanimate objects, (b) 3D facial scans performed on living persons.

The funnel plot showed asymmetry arising from 3 distinct points with different effect estimates (Figure 5). Regarding publication bias, Egger test results showed an intercept of 3.9 (95% CI 1.94-5.86, t=3.792, P=.004).
Figure 5

Funnel plot showing of publication bias assessment.

Global meta-analysis results of comparison of facial models obtained using mobile device–compatible face scanners versus professional face scanners. Subgroup meta-analysis results of comparison of facial models obtained using mobile device–compatible face scanners versus professional face scanners. (a) 3D facial scans performed on inanimate objects, (b) 3D facial scans performed on living persons. Funnel plot showing of publication bias assessment.

Discussion

Principal Findings

We aimed to investigate the accuracy of mobile device–compatible face scanners in facial digitization. Mean discrepancy values of the digitized face obtained using mobile device–compatible 3D facial scanners ranged from 0.34 to 1.40 mm in articles included in this systematic review. The meta-analysis revealed that mobile device–compatible 3D facial scanners were less accurate than their professional 3D counterparts. The reliability of a digital face scanner can be classified into 4 categories: highly reliable (deviation <1.0 mm), reliable (deviation 1.0 mm-1.5 mm), moderately reliable (deviation 1.5 mm-2.0 mm), and unreliable (deviation >2.0 mm) [46]. For clinical application, deviations <1.5 mm were considered acceptable [3,47,48]. Based on the classifications, mobile device–compatible 3D facial scanners were considered acceptable for clinical use even though their accuracies were lower than those of the professional 3D facial scanners. Amornvit et al [28] and Liu et al [43] reported that mobile device–compatible face scanners are comparable to professional 3D facial scanners when scanning simple and flat areas of the face such as the forehead, cheeks, and chin. However, scanning accuracy was relatively low when mobile device–compatible face scanners were used to capture complex facial regions, such as the external ears, eyelids, nostril, and teeth [28,44,45]. Higher inaccuracy was found in the facial areas with defects, depending on the depth of the defect [15]. Thus, careful consideration in accordance with the purpose and the person might be needed when using mobile device–compatible face scanners. In the preliminary stages, smartphone-based 3D scanners used a multiphotogrammetry approach that captured several photographs of the object from different views and matched common features in the photographs to establish a 3D model of the object by using dedicated smartphone software apps [15,45]. The resolution of a 3D image depended on the number of reconstructed polygons that were calculated by the software algorithm based on the resolution of the captured images [49]. The working principle is similar to that of professional stereophotogrammetry facial scanning systems; however, professional systems usually use digital single-lens reflex cameras that have higher pixel densities with better noise reduction software and higher ISO settings compared with those of smartphone cameras [50]. The accuracy of smartphone multiphotogrammetry in facial data acquisition was reported as 0.605 (SD 0.124) mm by Elbashti et al [15]. In another study by Ross et al [45], the mean discrepancy of scan data obtained using smartphones ranged from 0.9 mm to 1.0 mm, depending on the number of photographs taken during scanning. In recent years, infrared structured-light depth-sensing cameras have been incorporated in mobile devices to facilitate 3D optical scans [51]. 3D depth-sensing cameras work by the time-of-flight principle, measuring the time taken for light emitted by an illumination unit to travel to an object and back to the sensor array [52,53]. The 3D images are then reconstructed based on a depth map of the object and surroundings [54]. Although smartphone depth-sensing cameras share similar working principles with professional laser scanning systems, laser systems are more sensitive to depths because they are built with higher sensitivity sensors [15,23]. Amornvit et al [28] reported that the 3D depth-sensing sensor scanner of a smartphone is reliable in linear measurement at the frontal plane, but it has less accuracy in depth measurement compared with that of professional face scanners. Depth-sensing cameras can also be used separately and attached or plugged into smartphones, tablets, or laptop computers to acquire 3D scans [27,29,40-42,44]. Because the quality of facial scanning is also affected by the performance of compatible mobile devices when external depth-sensing cameras are used, the resulting accuracy might vary widely and should be evaluated for each combination of depth-sensing camera and mobile device. Subgroup meta-analysis showed that the accuracy of 3D facial scans performed on living persons was significantly different compared with those performed on inanimate objects. This result implies that the outcomes of in vitro or laboratory studies could be different from those obtained from people. Thus, based on the findings of this review, we recommend using living persons for related research on mobile device–compatible face scanning. Caution should be exercised when scanning the orbital, nasolabial, and oral regions on the face of a living person to minimize the discrepancies caused by motion artifacts [16,24]. Subconscious nose breathing, eye blinking, and lip twitching should also be carefully considered as these are the main sources of involuntary facial movements [16]. Ozsoy et al [17] reported that changes in facial expressions could affect the reproducibility and reliability of a scan, with the highest error values observed for a frightened facial expression and the lowest value observed for neutral facial expression. To reduce motion artifacts, the person should be instructed to maintain a neutral facial expression and avoid any head movement during image acquisition [55]. Another concern is that human faces contain complex skin textures, pores, freckles, scars, and wrinkles. Some artifacts or missing scan data appear as holes, originating from surfaces that are difficult to capture, such as eyebrows, eyelashes, and hairlines [29]. Small empty holes can be repaired using image processing software that uses neighboring areas that are morphologically similar; however, large defects can cause difficulties in the stitching process because of the lack of reference [24]. In addition, human faces vary in shape and are not perfectly symmetric, thus may appear different in different angles of view [56]. This phenomenon might cause some artifacts when the multiphotogrammetry approach is used because the 3D model of an object is reconstructed by matching common facial features in the captured photographs. A limitation of this review is that the review protocol was not preregistered on PROSPERO. Most included studies are not directly correlated with clinical treatment outcomes due to the difficulty of performing clinical studies to assess the accuracy of scanners. However, the findings of this review show great promise for the clinical use of mobile device–compatible face scanners. Another limitation of this systematic review is the small number of included studies. The limited number of studies show high heterogeneity and funnel plot asymmetry. Regarding publication bias, the Egger test result was significant (P=.004). Heterogeneity can cause funnel plot asymmetry when a correlation between intervention effects and study sizes is present [57]. Further examination was performed on the eligibility of a study that showed distinctly larger effect estimates [5], and we included the study [5] in the meta-analysis because it was conducted in an environment of a scanning intervention and was methodologically scientific. Although the inclusion of this study [5] increased the heterogeneity among studies and funnel plot asymmetry, the results were fundamentally attributed to a small number of articles [58]. All eligible papers included in the review were published between 2017 and 2020 due to the novelty of the research topic. A random-effects model is often used in meta-analyses for studies with heterogeneity. Random effects meta-analyses weigh studies more equally than fixed-effect analyses by incorporating the variance between studies [58]. Therefore, in this review, based on heterogeneity and funnel plot asymmetry, random-effects models were selected for global and subgroup analyses. Additional controlled in vitro and randomized clinical trials will be needed to reinforce the impact of review articles. Moreover, considering the rapid development of face scanning in the medical field, diverse investigations with newly developed devices and systems need to be continuously performed.

Conclusions

Overall, the accuracy of mobile device–compatible face scanners in 3D facial acquisition was not comparable to that of professional optical scanning systems, but it was still within the clinically acceptable range of <1.5 mm in dimensional deviation. There were significant differences between 3D facial scans performed on inanimate objects and living persons; thus, caution should be exercised when interpreting the results from studies conducted on inanimate objects.
  51 in total

Review 1.  Digital three-dimensional image fusion processes for planning and evaluating orthodontics and orthognathic surgery. A systematic review.

Authors:  Joanneke M Plooij; Thomas J J Maal; Piet Haers; Wilfred A Borstlap; Anne Marie Kuijpers-Jagtman; Stefaan J Bergé
Journal:  Int J Oral Maxillofac Surg       Date:  2010-11-20       Impact factor: 2.789

2.  In an empirical evaluation of the funnel plot, researchers could not visually identify publication bias.

Authors:  Norma Terrin; Christopher H Schmid; Joseph Lau
Journal:  J Clin Epidemiol       Date:  2005-09       Impact factor: 6.437

3.  Comparison of three-dimensional scanner systems for craniomaxillofacial imaging.

Authors:  Paul G M Knoops; Caroline A A Beaumont; Alessandro Borghi; Naiara Rodriguez-Florez; Richard W F Breakey; William Rodgers; Freida Angullia; N U Owase Jeelani; Silvia Schievano; David J Dunaway
Journal:  J Plast Reconstr Aesthet Surg       Date:  2017-01-07       Impact factor: 2.740

Review 4.  Three-dimensional esthetic analysis in treatment planning for implant-supported fixed prosthesis in the edentulous maxilla: review of the esthetics literature.

Authors:  Avinash S Bidra
Journal:  J Esthet Restor Dent       Date:  2011-05-12       Impact factor: 2.843

5.  Facial moulage fabrication using a two-stage poly (vinyl siloxane) impression.

Authors:  Abdullah S Alsiyabi; Glenn E Minsley
Journal:  J Prosthodont       Date:  2006 May-Jun       Impact factor: 2.752

6.  Reliability of optical devices for three-dimensional facial anatomy description: a systematic review and meta-analysis.

Authors:  D Gibelli; C Dolci; A Cappella; C Sforza
Journal:  Int J Oral Maxillofac Surg       Date:  2019-11-28       Impact factor: 2.789

7.  Digital dentures: A protocol based on intraoral scans.

Authors:  Lucio Lo Russo; Angelo Salamini; Giuseppe Troiano; Laura Guida
Journal:  J Prosthet Dent       Date:  2020-04-21       Impact factor: 3.426

Review 8.  Advancements in craniofacial prosthesis fabrication: A narrative review of holistic treatment.

Authors:  Hossein E Jazayeri; Steve Kang; Radi M Masri; Lauren Kuhn; Farahnaz Fahimipour; Rabecca Vanevenhoven; Geoffrey Thompson; Maryam Gheisarifar; Mohammadreza Tahriri; Lobat Tayebi
Journal:  J Adv Prosthodont       Date:  2018-12-19       Impact factor: 1.904

9.  The Accuracy of Digital Face Scans Obtained from 3D Scanners: An In Vitro Study.

Authors:  Pokpong Amornvit; Sasiwimol Sanohkan
Journal:  Int J Environ Res Public Health       Date:  2019-12-12       Impact factor: 3.390

10.  Facial landmark localization by curvature maps and profile analysis.

Authors:  Carsten Lippold; Xiang Liu; Kim Wangdo; Burkhard Drerup; Kristina Schreiber; Christian Kirschneck; Tatjana Moiseenko; Gholamreza Danesh
Journal:  Head Face Med       Date:  2014-12-08       Impact factor: 2.151

View more
  3 in total

1.  Accuracy of Portable Face-Scanning Devices for Obtaining Three-Dimensional Face Models: A Systematic Review and Meta-Analysis.

Authors:  Hang-Nga Mai; Jaeil Kim; Youn-Hee Choi; Du-Hyeong Lee
Journal:  Int J Environ Res Public Health       Date:  2020-12-25       Impact factor: 3.390

2.  The Effect of Perioral Scan and Artificial Skin Markers on the Accuracy of Virtual Dentofacial Integration: Stereophotogrammetry Versus Smartphone Three-Dimensional Face-Scanning.

Authors:  Hang-Nga Mai; Du-Hyeong Lee
Journal:  Int J Environ Res Public Health       Date:  2020-12-30       Impact factor: 3.390

3.  Effects of Artificial Extraoral Markers on Accuracy of Three-Dimensional Dentofacial Image Integration: Smartphone Face Scan versus Stereophotogrammetry.

Authors:  Hang-Nga Mai; Du-Hyeong Lee
Journal:  J Pers Med       Date:  2022-03-18
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.