Literature DB >> 35345376

Assessment of Videos on YouTubeTM about Nasopharyngeal Cancer in Terms of Accuracy, Reliability and Understandability.

Caglar Eker1, Ozgur Surmelioglu1, Sevinc Puren Yucel2, Muhammed Dagkiran1, Elvan Onan1.   

Abstract

INTRODUCTION: In the internet era we live in, it is very easy to access information. While this situation has positive effects for patients using the internet, it also brings some negative effects. The effects of the quality of YouTube™ videos on nasopharyngeal cancer were examined.
METHODS: "Nasopharyngeal Cancer" as search term was used to conduct a search on YouTube™. The 'Sort by' search filter was set at 'relevance', which is the default for YouTube™ searches. The first 250 results were reviewed and analyzed. After the videos were eliminated according to the exclusion criteria, 45 videos were evaluated by two authors. Video materials were categorized according to "video type" and "source of content". According to "video type" and "source of content" the videos were categorized into  two as educational and testimonial and three as medical institution, medical website, and individual users. After recording the features of all evaluated videos, accuracy score, audiovisual score, modified discern score, patient education materials assessment tool for audiovisual materials (PEMAT) score and usefulness score were determined for each video to evaluate the accuracy, reliability, and understandability of the videos.
RESULTS: The usefulness score, modified discern score, and accuracy score of the educational videos were significantly higher than testimonial videos (p<0.001 for all). Educational videos provided more useful and accurate video content than testimonial videos. In addition, it was also determined that the median PEMAT actionability score and audiovisual score of the individual group were statistically significantly lower than medical institutions and medical websites (p=0.001 and p<0.001, respectively). The videos provided by medical institutions, including universities, did not have a significant advantage over other groups in terms of accuracy, reliability, and usefulness.
CONCLUSION: Healthcare videos concerning nasopharyngeal cancer on YouTube™ are heterogeneous and are not peer reviewed. Therefore, medical professions on nasopharyngeal cancer need to upload more accurate, reliable and easy to understand videos onto online platforms such as YouTube™.

Entities:  

Keywords:  Internet; Nasopharyngeal cancer; Patient information; Youtube; accuracy

Mesh:

Year:  2022        PMID: 35345376      PMCID: PMC9360932          DOI: 10.31557/APJCP.2022.23.3.1023

Source DB:  PubMed          Journal:  Asian Pac J Cancer Prev        ISSN: 1513-7368


Introduction

Nasopharyngeal cancer (NPC) is a rare disease, especially in the western world, with and incidence of less than 1 per 100,000 person-years. However, in certain endemic regions such as Southern China, southeastern Asia, middle east and north Africa, this rate increases to 20 to 50 per 100,000 person-years (Zhang et al., 2015). According to 2018 data, approximately 129,000 new cases were detected in one year and approximately 73,000 deaths occurred due to the disease (Bray et al., 2018). The NPC is associated with multiple risk factors that include viral (Epstein-Barr Virus), genetic, and environmental factors. Out of these, dietary habits, exposure to harmful agents, and lifestyle probably explain the geographical differentiation. Early detection of nasopharyngeal cancer is difficult because it is not easy to examine and emerging symptoms mimic other more common diseases. It may also reveal very different symptoms such as lump in the neck, nasal obstruction, epistaxis, otitis media, hearing loss, facial numbness, and diplopia (Chen et al., 2019). Therefore, the patients usually experience misdiagnosis and delayed treatment. Ihe NPC is associated with poor prognosis, which is largely attributed to lack of awareness of the symptoms (Lee et al., 2019). Online health-related resources are becoming more popular and more effective in people’s decision-making processes. Ninety percent of American adults has internet access and %72 of these internet users state that they search for their medical condition on the internet (Fox, 2011). YouTube™ that is one of online resources is a free video streaming site with over 100 million daily viewers, with around 65,000 videos uploaded daily (YouTube Press Statistics, 2017). It is also a great resource for patient information with large number of contents, variety of videos, and practical uses. Whereas some of the videos are uploaded by professional sources such as medical professionals, most of the videos are based on personal experiences. This assortment of video content providers and lack of peer review process on YouTube™ has resulted in the broadcast of inadequate or misleading health information (Pandey et al., 2010). In addition, internet users who are not medical professionals may not be able to distinguish between genuine patient education videos and those made for commercial purposes. Therefore, it is important to emphasize that inappropriate and biased health-related information on the internet can influence patients to make irrational decisions (Balakrishnan et al., 2016). Therefore, there is a need to make sure that online users reach accurate, reliable, and understandable information sources. To the best of our knowledge, there has been no assessment of the accuracy, reliability, and intelligibility of the information presented on YouTube™ videos for NPC. We aimed to systematically analyze the videos that users are likely to encounter when searching YouTube™ for information about NPC. These videos were evaluated in terms of accuracy, reliability, intelligibility and actionability using measurement scales such as modified discern score, patient education materials assessment tool for audiovisual materials (PEMAT), NPC usefulness scoring systems, accuracy score, and audiovisual score.

Materials and Methods

The Cukurova University, Faculty of Medicine, Non-Invasive Clinical Research Ethics Committee approved the study, and the study was conducted in line with the Declaration of Helsinki. “Nasopharyngeal Cancer” as search term was used to conduct a search on YouTube (https://www.youtube.com/) on July 19, 2021. The ‘Sort by’ search filter was set at ‘relevance’, which is the default for YouTube searches. All researchers conducted the search by clearing their browser’s search history and disabling their location status to decrease bias. The first 250 results were reviewed and analyzed. Ads served by YouTube™ in search results were not counted. The videos were excluded based on exclusion criteria including non-English videos, non-relevant videos, duplicate videos, videos with no sound or subtitles, and meeting and webinar videos. Video analysis and data collections Two independent reviewers (C.E., O.S.) assessed and analyzed all videos separately. Firstly, identifying features of each video were recorded including URL; video name; video duration; upload date; the number of views, likes, dislikes and comments. View ratio (views/day) and likes ratio (likes * 100) / (likes * dislikes) were also calculated. As a result of our literature search for an index that evaluates both views and likes of videos, we chose to evaluate the popularity of videos using an index called Video Power Index (VPI), which is calculated with the following formula: like ratio*view ratio/100 (Erdem and Karaca, 2018). Then, each video was evaluated separately by applying all measurement scales. During the evaluation process of videos with measurement scales, some videos were rated differently by individual reviewers. Therefore, inter-rater reliability was assessed for each measurement scales. Three possible answers were given to the “source of content” category. If responses by reviewers differed, the video was considered as “unidentified source”. In terms of the accuracy score, audiovisual score, modified discern score, PEMAT score (understandability and actionability) and usefulness score, the median value would be taken into account if results found by reviewers differed. Video materials were categorized according to “video type” and “source of content”. It was divided into two; 1) educational and 2) testimonial videos according to the purpose of production. In terms of source of content, the videos were in three categories; 1) medical institution (the official account of a university, hospital and an individual who works in a university and hospital), 2) medical website (healthcare related youtube channels or medical charity foundations) and 3) individual users (a private account with no affiliation to an institution or a university). The features and measurement scale results of the groups within the same category were compared with each other. Audio-visual quality scores and accuracy scores were determined after the videos were classified. Audio-visual quality scores were given out of four using the following ratings: 0 = impossible to view, 1 = poor-blurred, out of focus; 2 = moderate-non-professional editing; 3 = excellent quality-clear, professional editing. Accuracy scores were also given out of four: 0 = misleading and largely false; 1 = poor-easily identifiable misinformation; 2 = moderate-some oversimplification, general correct information; 3 = excellent-professional level, extremely accurate (Enver et al., 2020). Reliability of video information was determined using the modified discern score, as originally defined by Singh (2012). The modified discern score assesses clarity, credibility, bias, reference reinforcement, and areas of uncertainty, specifically for information in YouTube™ videos (Supplement 1). One point is given for each criterion, with five points indicating the highest reliability. The PEMAT is a systematic method developed to select printable and audiovisual patient education materials, which are easier to understand and easier to act on. We used the version for audiovisual materials, which consists of thirteen items measuring understandability and four items measuring actionability (Supplement 2). The PEMAT provides two scores for each material, one for understandability and a separate score for actionability. Every item has 1 point (Agree) or 0 points (Disagree) and not applicable (N/A - not included the calculation). The total scores of the material on understandability and actionability items were summed up separately. The total score was divided by the total possible score- i.e. the number of items for which the material was rated, excluding items scored as not applicable (N/A). The result was multiplied by 100 to obtain a percentage (%) for understandability and actionability. There was no set cut-off value for the scores (PEMAT, 2020). Since there is no current usefulness index for nasopharyngeal cancer, a new scoring system was developed for this study to review the reliability of the videos (Supplement 3). In this scoring system, we have listed the main information that a reasonable doctor should give to the patient about NPC. We adapted the scoring systems made in previous studies for NPC in order to evaluate the YouTube™ videos as well as reviewing the current literature (Ben-Ami et al., 2021; Bossi et al., 2021). The usefulness score consisted of a total of 12 items; Ebstein Barr virus, nutrition and habits (Salted fish, smoking, and alcohol), geographical distribution, age, gender and genetics, nasal obstruction, epistaxis, neck mass, additional symptoms (otitis media, sinusitis, diplopia, and facial paresthesia), diagnosis, treatment options, and progress of the disease. The videos were checked in terms of the twelve items being mentioned one by one, and each mentioned item was worth one point. According to this, scores between 10 and 12, 7 and 9, 4 and 6, and 0 and 3 were categorized as very useful, useful, slightly useful, and poor respectively. The usefulness scores of the videos were calculated, categorized, and compared. In addition, the videos were divided into three according to the usefulness category; 1) useful group, 2) slightly useful group, and 3) poor group. The features and measurement scale results of the groups were compared with each other. Statistical analysis Categorical variables were expressed as numbers and percentages, whereas continuous variables were summarized as mean and standard deviation and as median and minimum-maximum. Chi-square test was used to compare categorical variables between the groups. The normality of distribution for continuous variables was confirmed with the Shapiro Wilk test. For comparison of continuous variables between two groups, Mann Whitney U test was used. For non-normal distributed data, Kruskal Wallis test was used to compare more than two groups and Bonferroni adjusted Mann Whitney U test was used for multiple comparisons of groups. To evaluate the correlations between measurements, Spearman Rank Correlation Coefficient was used. Inter-rater reliability was evaluated with Fleiss kappa and Spearman Rank Correlation coefficient. All analyses were performed using IBM SPSS Statistics Version 20.0 statistical software package. The statistical level of significance for all tests was considered to be 0.05.

Results

The first 250 results of the search were reviewed and analyzed. Videos that did not meet the criteria (n:178) and could not be categorized due to “unidentified source” (n:27) were excluded from the study. A total of 45 videos were included in the study. After classification by type, the educational and testimonial groups consisted of 22 and 23 videos respectively. After the classification by source of content, the healthcare channel, university and individual groups consisted of 18, 12, and 15 videos respectively. In addition, videos were classified according to the NPC usefulness scoring system and no very useful video was found. In this case, useful, slightly useful, and poor video groups consisted of 8, 11, and 26 videos respectively. The mean length of the videos was 180 second (s) (range from 25 s to 1912 s). The mean number of views for nasopharyngeal cancer-related videos was 730 (range: 2 – 69935 views); each video was viewed for a mean of 0,3 views/day (range: 0 – 45 views/day). The inter-rater reliability of the scores is listed in Table 1. The Fleiss kappa coefficients for the scores ranged from 0.547 to 0.961 and the Spearman correlation coefficients were ranged from 0.887 to 0.982, showing statistically significant inter-rater reliability. Inter-rater reliability was statistically significant and almost excellent for all scores (p<0.001). This result means that the two reviewers agreed on all scores about the YouTube™ videos in terms of NPC.
Table 1

Assessment of Interrater Reliability for Measurement Scales

Reliability measureDiscern Score(0-4)PEMAT score(Understandability)PEMAT score(Actionability)Accuracy score (0-3)Audiovisual score (0-3)Usefulness score (0-10)Usefulness(0-2)
Fleiss Kappa inter-rater reliability (p-value)0.8 (<0.001)--0.880 (<0.001)0.675 (<0.001)0.547 (<0.001)0.961 (<0.001)
Spearman’s correlation coefficient (p-value)0.962 (<0.001)0.968 (<0.001)0.891 (<0.001)0.926 (<0.001)0.887 (<0.001)0.969 (<0.001)0.982 (<0.001)

PEMAT, Patient education materials assessment tool for audiovisual materials

Descriptive statistics of included videos according to video type are given in Table 2. While there was no difference in the median number of dislike and comment (p=0.586, p=0.547 respectively), the median number of likes and VPI values of the testimonial group (7 and 0.6, respectively) were higher than the educational group (3.5 and 0.2 respectively). However, this difference was not statistically significant (p=0.501 and p=0.776) (Figure 1a-b). The median usefulness score, modified discern score, and accuracy score of the educational videos were significantly higher than testimonial videos (p<0.001 for all), and their distribution is shown in Figure 1c-e. Although, the median audiovisual score, PEMAT understandability, and actionability index values of educational videos were higher compared to testimonial videos, this difference was not significant (p=0.127, p=0.114, and p=0.071 respectively).
Table 2

Summary Statistics of Included Videos According to Video Type

Video type p
Educational(n=22)Testimontial(n=23)
Video durationsa(s)152 (35-878)278 (25-1912)0.115
Upload dateda988.5 (236-4359)1173 (86-3099)0.601
Views numbera550.5 (2-69935)761 (27-67595)0.642
Views Daya0.2 (0-45)0.6 (0-26.7)0.768
Likes numbera3.5 (0-707)7 (0-645)0.501
Dislike numbera0 (0-20)0 (0-15)0.586
Comment numbera0 (0-96)0 (0-140)0.547
Like Ratioa94.5 (0-100)98 (0-100)0.317
VPIa0.2 (0-43.6)0.6 (0-25.9)0.776
Discern Scorea3 (0-4)1 (0-2)<0.001
"PEMAT scorea (Understandability)"63.5 (0-92)50 (11-85)0.114
"PEMAT scorea (Actionability)"67 (0-100)33 (0-100)0.071
Accuracy scorea2 (0-3)1 (0-2)<0.001
Audiovisual scorea2 (0-3)1 (1-3)0.127
Usefulness scorea4.5 (0-10)2 (0-6)<0.001
Usefulnessb0.001
Poor7 (31.8)19 (82.6)
Slightly useful7 (31.8)4 (17.4)
Useful8 (36.4)0 (0.0)
Source of contentb0.322
Healthcare channel/charity11 (50.0)7 (30.4)
University/Hospital4 (18.2)8 (34.8)
Individual users7 (31.8)8 (34.8)

a, Data were expressed as median(min-max); b, n(%); s, second; VPI, Video power indeks; PEMAT, Patient education materials assessment tool for audiovisual materials

Figure 1

The Distributions of Number of Likes (a), VPI score (b), Usefulness score (c), Modified Discern score (d) and Accuracy score (e) according to video type (educational vs. testimonial)

The distributions of PEMAT actionability score and audiovisual score were not similar between source of content groups (p=0.001 and p<0.001, respectively) (Figure 2a-b). As a result of pairwise comparisons, it was determined that the median PEMAT actionability score and audiovisual score of the individual group were lower than the other groups. Additionally, usefulness score, accuracy score, and modified discern score were similar among the source of context groups (p=0.647, p=0.424, and p=0.155 respectively) (Table 3) (Figure 2c-e).
Figure 2

Distributions of PEMAT Score (actionability) (a), Audiovisual score (b), Usefulness score (c), Accuracy score (d) and Modified Discern score () according to source of content

Table 3

Summary Statistics of Included Videos According to Source of Content

Source of Contentp
Healthcare channel/charity (n=18)University/Hospital (n=12)Individual users (n=15)
Video durationsa(s)137 (33-1912)217 (75-1365)180 (25-1595)0.546
Upload dateda1028.5 (86-4359)999.5 (274-3099)1117 (276-2747)0.910
Views numbera694 (10-69935)434.5 (9-23740)730 (2-67595)0.710
Views Daya0.7 (0-45)0.2 (0-28.7)0.3 (0-26.7)0.343
Likes numbera5.5 (0-707)4.5 (0-139)5 (0-645)0.786
Dislike numbera0 (0-20)0 (0-8)0 (0-15)0.706
Comment numbera0 (0-96)0 (0-26)0 (0-140)0.728
Like Ratioa96.5 (0-100)92 (0-100)97 (0-100)0.742
VPIa0.7 (0-43.6)0.2 (0-27)0.3 (0-25.9)0.605
Discern Scorea2 (0-4)1 (0-4)1 (0-3)0.155
"PEMAT scorea (Understandability)"65 (0-92)55 (20-90)40 (11-81)0.063
"PEMAT scorea (Actionability)"67 (0-100)67 (0-100)33 (0-67) Ψ, ɸ0.001
Accuracy scorea1.5 (0-3)1 (0-3)1 (0-3)0.424
Audiovisual scorea2 (0-3)2 (1-3)1 (0-2) Ψ, ɸ<0.001
Usefulness scorea3.5 (0-9)3 (0-10)2 (0-8)0.647
Usefulnessb0.914
Poor9 (50.0)7 (58.3)10 (66.7)
Slightly useful5 (27.8)3 (25.0)3 (20.0)
Useful4 (22.2)2 (16.7)2 (13.3)
Video typeb0.322
Educational111 (61.1)4 (33.3)7 (46.7)
Testimonial7 (38.9)8 (66.7)8 (53.3)

a, Data were expressed as median(min-max); b, n(%); Ψ p<0.05 compared with healthcare channel/charity group; ɸ p<0.05 compared with university/hospital group; s, second; VPI, Video power indeks; PEMAT, Patient education materials assessment tool for audiovisual materials

Descriptive statistics of included videos according to usefulness are given in Supplement 4. The median modified discern and accuracy score of the poor group was statistically significantly lower than slightly useful and useful groups (both p<0.001). The PEMAT understandability score of the poor group was significantly lower than the useful group (p=0.031). Similarly, PEMAT actionability score of the poor group was statistically lower than the slightly useful group (p=0.036). We also examined the correlations between the suggested usefulness score and other video characteristics (Supplement 5). Correlation analysis demonstrated that usefulness score significantly and positively correlated with accuracy score (r=0.794, p<0.001 and r=0.747, and p<0.001 respectively). The Distributions of Number of Likes (a), VPI score (b), Usefulness score (c), Modified Discern score (d) and Accuracy score (e) according to video type (educational vs. testimonial) Assessment of Interrater Reliability for Measurement Scales PEMAT, Patient education materials assessment tool for audiovisual materials Summary Statistics of Included Videos According to Video Type a, Data were expressed as median(min-max); b, n(%); s, second; VPI, Video power indeks; PEMAT, Patient education materials assessment tool for audiovisual materials Summary Statistics of Included Videos According to Source of Content a, Data were expressed as median(min-max); b, n(%); Ψ p<0.05 compared with healthcare channel/charity group; ɸ p<0.05 compared with university/hospital group; s, second; VPI, Video power indeks; PEMAT, Patient education materials assessment tool for audiovisual materials Distributions of PEMAT Score (actionability) (a), Audiovisual score (b), Usefulness score (c), Accuracy score (d) and Modified Discern score () according to source of content

Discussion

In today age of technology, it is very easy to access information and this condition causes specific problems. There is an incredible amount of information pollution on the internet. If the user is not an expert on the subject they research, they cannot determine whether the information is correct. YouTube™ is one of these sources of information. Similar to other popular social media websites, YouTube™ also shares materials for free and allows any registered user to upload health-related, non-peer-reviewed videos, and the information contained in these videos is not checked. This makes YouTube™ vulnerable to posting fake and potentially dangerous videos that are not backed by solid scientific evidence. In our study, regardless of the video source, the median accuracy score was 1/3, the mean usefulness score was 3/12, and the mean modified discern score was 1/5. Especially in testimonial videos, these values were found to be lower. Additionally, symptoms and treatment options were not mentioned in most of the videos. Drozd’s (2018) review highlighted that most health-related YouTube videos present false and unreliable information. Given the results, we consider that YouTube™ is not a reliable and accurate source for NPC. It was found that only 26% of them were medical professionals in official medical institutions. The accuracy and reliability of the information given by the rest is highly suspicious. The heterogeneous and uncontrolled misinformation on YouTube has been previously reviewed by Hassona et al., (2016) in their own study on oral cavity cancer. Due to this information pollution, patients can be misled and make wrong decisions even on a serious and sensitive issue such as cancer. The videos available on YouTube™ vary in terms of quality, accuracy, and content as these materials are very heterogeneous. Enver et al., (2020) showed that the audiovisual quality scores and the accuracy score of the university group were statistically significantly higher than the health-care channels, individual users, and TV channel/news groups. In the same study, authors also indicated that videos uploaded by university-affiliated accounts were more accurate, more trustworthy, and more professionally recorded and edited. In our study, we also observed that audiovisual quality score of the individual group were lower than the universities and health-care groups. However, there were no significant difference between university/medical institution and health-care groups. Additionally, accuracy scores between these three groups did not have significant differences. Therefore, our findings indicated that university/medical institution group was not more reliable, more accurate, and more professional than health-care group for NPC. In the study by Fode et al., (2020) to evaluate videos content on erectile disfunction, the median PEMAT understandability score and the median PEMAT actionability score were equal to 100%. In our study, the PEMAT score differed according to the source from which the videos were uploaded. The PEMAT actionability scores of the videos uploaded by university/medical institution and healthcare groups were statistically significantly higher compared to individual group. There were differences between these groups about PEMAT understandability score; however, these differences were not statistically significant. The fact that most of the videos in the university/medical institution group consisted of patient testimonial videos (%66.7) may explain that these videos are more understandable and actionable. However, the accuracy and reliability values were not better than the other groups. We think that these videos are not helpful for patients to make the right choice. Therefore, it is clear that there is a need for official medical institutions to upload more accurate and reliable content on NPC. Videos from a different perspective as educational and testimonial were examined. Although there was no difference between the audiovisual quality scores of the educational and testimonial videos, the accuracy and usefulness scores of the educational group were statistically higher than testimonial group. It was also found that modified discern score of educational group was significantly higher than testimonial group. Educational videos were more reliable, more accurate, and useful than testimonial videos. Moreover, there was no significant difference between the two groups in terms of VPI score, number of views, and likes. Radonjic et al., (2020) demonstrated that videos with the highest VPI values, a measure of video popularity, had the lowest reliability scores. Several studies have also indicated a paradoxically high popularity among patient testimonials videos, which do not cite evidence-based research for their claims (Fischer et al., 2013; Erdem and Karaca, 2018; Gokcen and Gumussuyu, 2019). Similarly, in our study, even though 48% of the videos were delivered by physicians for educational purposes, the VPI scores on these videos were not significantly higher than the testimonial videos. It is clear that more useful videos are more accurate, more reliable, more understandable, and more actionable for treatment decision-making process of patient. It was demonstrated that there were statistically significant correlations between usefulness score and accuracy score, modified discern score, and PEMAT scores (understandability and actionability). This situation may reveal that the usefulness score system developed by the researchers can be used in parallel with other scoring systems in the literature. However, it was also demonstrated that there were no statistically significantly correlations between usefulness score and views number, viewing rate, and VPI score. Studies have also achieved similar findings on videos about oral cavity cancer and laryngeal cancer respectively (Hassona et al., 2016; Enver et al., 2020). Thus, these findings show that useful, accurate, reliable healthcare related videos on YouTube™ are not viewed enough. In addition, these findings mean that videos that provide misleading and incomplete information are watched at least as often as useful educational videos. The first limitation of the present study is the pretty dynamic structure of YouTube™. The videos are continuously uploaded and deleted. A video you watch today may be deleted the day after. Future studies assessing healthcare related videos on YouTube™ can solve this issue by utilizing different methods. Second, we analyzed only English videos, while most of the regions where NPC is more common are not English speakers. Extending the coverage of languages especially southeastern Asia languages, it is possible to obtain more accurate results. Lastly, YouTube™ is not the only online platform which people can prefer to access videos. Planning more comprehensive studies to include other video platforms would provide more reliable results. In conclusion, YouTube™ is a place that patients can research information about their diseases and treatment options. Healthcare videos on YouTube™ are heterogeneous and are not peer reviewed. People view these videos without making a distinction between poor and useful content and they do not know how to make such a distinction. Therefore, these misleading videos can lead viewers into wrong decisions. In our age, social media is an obvious unneglectable source and otorhinolaryngologists need to upload more accurate, reliable, and easy to understand videos onto platforms such as YouTube™.

Author Contribution Statement

Conceptualization, C.E. and O.S.; Design, C.E., O.S. and M.D.; Supervision, C.E., O.S., M.D. and E.O.; Data Collection and/or Processing, C.E., O.S., S.P.Y.K., M.D. and E.O.; Analysis and/or Interpretation, C.E., O.S. and S.P.Y.K.; Literature Search, C.E., O.S. and S.P.Y.K.; Writing Manuscript, C.E., O.S., S.P.Y.K., M.D. and E.O.; Critical Review, C.E., O.S., S.P.Y.K., M.D. and E.O.
  17 in total

1.  YouTube as a source of information on mouth (oral) cancer.

Authors:  Y Hassona; D Taimeh; A Marahleh; C Scully
Journal:  Oral Dis       Date:  2016-01-21       Impact factor: 3.511

2.  Educational quality of YouTube videos on knee arthrocentesis.

Authors:  Jonas Fischer; Jeroen Geurts; Victor Valderrabano; Thomas Hügle
Journal:  J Clin Rheumatol       Date:  2013-10       Impact factor: 3.517

3.  YouTube as a source of information on the H1N1 influenza pandemic.

Authors:  Ambarish Pandey; Nivedita Patni; Mansher Singh; Akshay Sood; Gayatri Singh
Journal:  Am J Prev Med       Date:  2010-03       Impact factor: 5.043

Review 4.  Nasopharyngeal carcinoma.

Authors:  Yu-Pei Chen; Anthony T C Chan; Quynh-Thu Le; Pierre Blanchard; Ying Sun; Jun Ma
Journal:  Lancet       Date:  2019-06-06       Impact factor: 79.321

5.  A Quality Analysis of Disc Herniation Videos on YouTube.

Authors:  H Bahadir Gokcen; Gurkan Gumussuyu
Journal:  World Neurosurg       Date:  2019-02-02       Impact factor: 2.104

6.  YouTube™ as an information source for larynx cancer: a systematic review of video content.

Authors:  Necati Enver; Can Doruk; Hakan Kara; Ece Gürol; Sefa Incaz; Ulker Mamadova
Journal:  Eur Arch Otorhinolaryngol       Date:  2020-03-16       Impact factor: 2.503

7.  Readability and Understandability of Online Vocal Cord Paralysis Materials.

Authors:  Vini Balakrishnan; Zachariah Chandy; Amy Hseih; Thanh-Lan Bui; Sunil P Verma
Journal:  Otolaryngol Head Neck Surg       Date:  2016-02-09       Impact factor: 3.497

8.  Evaluating the Accuracy and Quality of the Information in Kyphosis Videos Shared on YouTube.

Authors:  Mehmet Nuri Erdem; Sinan Karaca
Journal:  Spine (Phila Pa 1976)       Date:  2018-11-15       Impact factor: 3.468

9.  Incidence trend of nasopharyngeal carcinoma from 1987 to 2011 in Sihui County, Guangdong Province, South China: an age-period-cohort analysis.

Authors:  Li-Fang Zhang; Yan-Hua Li; Shang-Hang Xie; Wei Ling; Sui-Hong Chen; Qing Liu; Qi-Hong Huang; Su-Mei Cao
Journal:  Chin J Cancer       Date:  2015-05-14

10.  Quality of Information in YouTube Videos on Erectile Dysfunction.

Authors:  Mikkel Fode; Alexander B Nolsøe; Frederik M Jacobsen; Giorgio Ivan Russo; Peter B Østergren; Christian Fuglesang S Jensen; Maarten Albersen; Paolo Capogrosso; Jens Sønksen
Journal:  Sex Med       Date:  2020-06-24       Impact factor: 2.491

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.