Literature DB >> 33546572

Interactions between emotional and cognitive engagement with science on YouTube.

Ilana Dubovi1, Iris Tabak2.   

Abstract

This study aimed to map and characterize public engagement with science on YouTube. A two-part study was conducted. First, we collected and quantitatively analyzed trending videos on YouTube to evaluate the magnitude of public interaction with science content. Then, we assessed actual, rather than self-reports of, media interactions with science-related YouTube trending videos. We tested associations between behavioral engagement of viewing, liking, disliking or commenting, and emotional and cognitive engagement. Our findings affirm that science content attracts high public interest and that emotional and cognitive engagement with science on social media are distinct, but interrelated. We show that regardless of the valence of emotional engagement, emotion is linked to greater behavioral engagement of posting comments and to greater cognitive engagement of argumentative deliberation. Therefore, our findings suggest that social media interactions, which tend to evoke emotional responses, are a promising means of advancing person-to-person engagement with science.

Entities:  

Keywords:  YouTube; argumentation; behavioral engagement; cognitive engagement; emotional engagement; public engagement in science; sentiment analysis; social media

Mesh:

Year:  2021        PMID: 33546572      PMCID: PMC8314998          DOI: 10.1177/0963662521990848

Source DB:  PubMed          Journal:  Public Underst Sci        ISSN: 0963-6625


Social media is gaining momentum as both a context for public engagement with science, and a target for research on public engagement with science (Hargittai et al., 2018; Huber et al., 2019; Kahle et al., 2016; Metag, 2020). In fact, social media may offer alternatives to people who want to engage with science but have reservations concerning traditional or print media (Metag, 2020). However, little is known about the nature of public engagement with science through social media (Hargittai et al., 2018; Kahle et al., 2016; Metag, 2020). In addition, the term engagement connotes multiple meanings and there is a need for greater precision and conceptualization of this term (Hargittai et al., 2018; Ji et al., 2019; Jünger and Fähnrich, 2020; McCallie et al., 2009). In this article, we examine public engagement with science on YouTube, focusing on naturally occurring online behaviors, employing a more refined, multidimensional, conceptualization of engagement. The term “engagement with science” is used to mark an active and reciprocal interaction between public and scientific content through interaction with media, other members of the public or formal scientific entities, such as scientists, or governmental research institutes (Hargittai et al., 2018; Khan, 2017; McCallie et al., 2009). Yet, “engagement” invariably denotes a broad range of activities and interactions from consumption such as reading or viewing of scientific information, to interaction, such as deliberation over the merit, applicability or moral implications of specific scientific findings, to public action, such as influencing policy (McCallie et al., 2009). Overall, the term “engagement” is underspecified and equivocal (Hargittai et al., 2018; Ji et al., 2019; Jünger and Fähnrich, 2020; McCallie et al., 2009). In this article, we synthesize literature from communication, psychology, and education to articulate a conceptualization of engagement on social media. Less is known about public interaction with science in the context of social media than in the context of print or older media such as radio, television, and standard websites. In the context of social media, little is known, for example, about the object of interactions, the extent of interactions, the nature of interactions, and how interactions manifest in different social media platforms. While large-scale surveys are essential for identifying public science information sources and attitudes toward different sources (Funk et al., 2017; National Science Board, 2018), they do not necessarily measure whether people skim or immerse in these online sources, agree or disagree with the information or what emotional responses they evoke in the real world. Yet, this knowledge is key for understanding the role of new media in supporting an agentive public. In this article, we apply a conceptual framework of engagement on social media to analyses of user interactions and to semantic analyses of comments on science-related YouTube channels, in order to gain insight into such understanding. We focus on YouTube for a few reasons. First, as one of the top three platforms for social media and for broadcasting content (Alexa Internet Inc., 2020; Perrin and Anderson, 2019), it merits attention and study. Second, the myriad of content options and scale of use offer potential for studying a diverse audience with respect to public engagement with science. Finally, the post-video comment space presents a context where debate and deliberation can be observed and measured (Dubovi and Tabak, 2020), yet this comment space is understudied (Murthy and Sharma, 2019). Our two research questions are (1) what is the proportion of engagement with science within top trending videos on YouTube? and (2) what characterizes behavioral, cognitive, and emotional engagement with scientific content?

1. Literature review

Public participation in science on social media

Investigating Internet and social media users’ comments on science-related sites and discussions can provide a window into the interests, attitudes, thoughts, and deliberations of a myriad of people who use these sites world-wide. For example, Kahle et al. (2016) found that visiting a site correlated with curiosity and with commenting, they further found that commenting corresponded with controversy and that arguments were associated with longer threads. This suggests that comments might reflect a deeper level, or different form or quality of engagement. Considerable research on user comments examines the lifespan and flow of ideas, tracking elements such as retweeting, or one comment taking up ideas from another comment (Bhattacharya et al., 2014; Hine, 2014; Salathé and Khandelwal, 2011). Scientists and science communicators impose a frame (Jünger and Fähnrich, 2020; Nisbet and Mooney, 2007; Tabak, 2016) on the article or video that they publish and post. That is, the content that they present conveys not only facts, but also attitudes and positions, even if these are not expressed explicitly (Chong and Druckman, 2007; Entman, 1993). Therefore, one line of research examining the content of user comments studies whether the tone, position expressed or topic of user comments correspond with those of the focal news article or video. For example, Tian (2010) found that how a video is framed in its written description on YouTube altered the valence of the comments about that video (Tian and Yoo, 2020). Similarly, in a variety of videos on genetically modified food, different sources, such as advocacy groups or scientists, evinced different tones, such as positive, negative or neutral, and user comments similarly varied by source but generally reflected the tone of the video (Chi et al., 2018). Unlike sentiments and tone, there was no clear pattern in the topics that users raised in their comments, some topics were similar to the topics raised in the video, and some topics were different (Kahle et al., 2016; Knowles and Wilkinson, 2015; Tian, 2010; Zhang, 2016). It is possible that similarity in tone and sentiment is a result of a process of self-selection, where people choose to view and comment on videos that correspond with their views. However, expressions of disagreement are sometimes associated with extensive comments (Dubovi and Tabak, 2020), suggesting that disagreement as well as agreement draw viewing and commenting. Therefore, the relationship between video content, and tone and sentiment of comments is complex, and comment analysis alone does not provide sufficient information to determine whether video content affects comment sentiment. Increasingly, scholars have become interested in how people discuss science on social media and whether these discussions help to create a deeper understanding of how the public uses and constructs scientific knowledge. One consistent finding is that user comments include many references to personal experiences, and that these personal experiences are considered credible sources of evidence by the commentators as well as by other users (Kahle et al., 2016; Knowles and Wilkinson, 2015; Len-Ríos et al., 2014; Morphett et al., 2020; Zhang, 2016). People are unlikely to cite sources in support of their claims (Zhang, 2016). This does not necessarily mean that the public sees no need for empirical evidence, rather, it may be a function of perceiving social media’s communicative genre as informal conversation that does not call for providing evidentiary support for claims. In fact, online comments often question and critique scientific methods and interpretations (Len-Ríos et al., 2014; Orr and Baram-Tsabari, 2018). When social media users do cite sources, these sources vary considerably from personal experiences, to published research, to news articles, to Google searches, to the Bible or to quotes from well-known individuals (Zhang, 2016). One characteristic that distinguishes public discussions on social media from formal scientific communications is in the expression of emotion in relation to scientific topics. Participants express fear (Orr and Baram-Tsabari, 2018), anger (Knowles and Wilkinson, 2015), and other emotions in their comments. Thus, their interactions with scientific content are both cognitive and emotional. Media consumption can have cognitive, affective, attitudinal, and behavioral effects, and these different effect-dimensions can interact (Maier et al., 2014). For example, affective reactions to science news media can influence judgments such as risk assessment (Maier et al., 2014). Therefore, more attention should be given to the ways in which these different dimensions play out in social media’s role in public engagement with science, because this may affect how people apply scientific knowledge, and it may also affect the trust people place in public scientific research.

The construct of engagement and how it manifests in social media

Engagement is a multidimensional construct that refers to an individual’s active involvement in an activity (Christenson et al., 2012). Operationalizations of engagement have been offered from a variety of theoretical and practical approaches (Christenson et al., 2012; D’Mello et al., 2017). Grounded in communication, psychological, and educational research, we synthesize literature on the construct of engagement, articulating three types of engagement: behavioral, emotional, and cognitive (D’Mello et al., 2017; Fredricks et al., 2004; Sinatra et al., 2015; Skinner, 2016).

Behavioral engagement on social media

Behavioral engagement is the outward manifestation of involvement. In educational contexts, it refers to students’ involvement in observable behavior directly related to the learning process, starting with attendance and continuing along the spectrum to attentiveness, compliance, effort expenditure, concentration, focus, persistence, hard work, and taking the initiative in academic tasks (Sinatra et al., 2015; Skinner and Belmont, 1993). Educational research shows a robust link between behavioral engagement and achievement (Ladd and Dinella, 2009; Luo et al., 2009; Wang and Eccles, 2012). Online behavioral engagement on various social media platforms is typically expressed symbolically through actions such as liking, commenting, and sharing. On YouTube, it also includes uploading and sharing videos, viewing videos, and reading comments. Users may choose to remain passive by simply consuming content, or play an active role by participating in various interactions, and even repurpose content to fit their needs (Khan, 2017). Posting comments signifies a higher form of engagement. Comments can reveal and expose personal meaning and resources, which in turn can transform user–user interaction into dialogue (Kahle et al., 2016; Muntinga et al., 2011; Shao, 2009).

Emotional engagement on social media

A number of theoretical traditions specify different components of emotional engagement (Christenson et al., 2012; D’Mello et al., 2017). Theories of motivation, including self-determination theory, and the control-value theory of academic emotions (Deci and Ryan, 1985; Pekrun and Linnenbrink-Garcia, 2012), emphasize how positive and negative emotions that are generated by learning situations influence involvement in learning activities. Positive emotions include enthusiasm, interest, satisfaction, and enjoyment (Renninger and Bachrach, 2015). Disaffected emotional components include boredom, anxiety, and frustration (Skinner and Belmont, 1993). Since behavioral and emotional engagement in the classroom are tightly coupled, both have been found to be strong predictors of grades and achievement test scores (Connell et al., 1995). On social media, emotional engagement is mostly inferred from comment text through computational techniques such as sentiment analysis (Mohammad, 2016) that typically identify the components and intensity of emotional arousal discussed above (Huang and Grant, 2020). In these techniques, a lexicon for a set of possible emotions (e.g. joy, anger, sadness), depicting a collection of words directly (e.g. delighted as an index of joy) and indirectly (e.g. shout as an index of anger) related to the set of emotions is used to analyze and compute instances of words from the text. Other approaches infer emotions through content analysis or other qualitative analyses that focus more closely on the content and meaning of texts in context, such as in tweets or in the context of a conversation thread (Gaspar et al., 2016). Sentiment analysis offers greater consistency and reliability because it is not dependent on case-by-case judgment and interpretation; it is also resource-lean and efficient. However, unlike qualitative approaches, it may overlook nuances of meaning that arise from the ways in which particular words are used in context. Research that employs such sentiment analysis reveals associations between commenting patterns and sentiments. Thelwall et al. (2012) report that while user comments are typically mildly to moderately positive, 35% of comments contain some negativity. They explain that although negative sentiment was uncommon, it was more prevalent in comments for videos attracting many comments; conversely, positive sentiment was disproportionately common in videos attracting few comments. Thus, it seems that negativity can drive commenting, perhaps partly through long-running, acrimonious, comment-based discussions. Similarly, a large-scale study using 756 popular queries to generate 67,290 videos with 6.1 million comments has investigated the role of sentiment in categories and the ratings of comments (i.e. the extent to which YouTube users rate a comment as good or bad), finding that ratings were predominantly positive (Siersdorfer et al., 2010). This study also found that negative comments tended to be disliked and positive comments tended to be liked, similar to the relationships between behavioral and emotional engagement found in learning contexts.

Cognitive engagement on social media

Cognitive engagement is more difficult to define precisely. The “problem that plagues construct definition here is lack of agreement among scholars about how cognitive engagement should be operationalized” (Sinatra et al., 2015). A widely used definition of cognitive engagement is psychological investment, which refers to students’ mental orientation during learning activities, such as an orientation toward deeper understanding and growth, a preference for challenge, and use of self-regulatory or coping strategies. Students become psychologically invested when they expend cognitive effort in order to understand, go beyond the requirement of the activity, use flexible problem-solving, and choose challenging tasks. User post-video comments on YouTube can reflect this approach to cognitive engagement because users go beyond the basic task requirements of viewing to posting. Of course, the comments themselves can be relatively flippant comments, such as “great,” “thanks for posting” or “very timely.” Comment contents can further reflect cognitive engagement in different ways. One form of such engagement, often measured through self-report surveys (Jensen, 2011; Oeldorf-Hirsch, 2018; Perse, 1990), is elaborating on the information provided in the original post. The asynchronous nature of post-video comments on YouTube and other digital media supports deeper engagement by giving participants more time to think, reflect, and seek information before they contribute to the discussion (Lucas et al., 2014). Another form of cognitive engagement on social media is argumentation (Asterhan and Hever, 2015; Dubovi and Tabak, 2020; Shapiro and Park, 2015, 2018; Tsovaltzi et al., 2015), where users respond to each other’s comments by further elaborating, challenging claims in the post or other comments, or by providing evidence to support or refute claims. In this way, cognitive engagement is both an individual and collaborative accomplishment.

2. Methods

Research design

We conducted a two-part study. Part 1 included collection and quantitative analysis of YouTube trending videos to evaluate the magnitude of public interaction with science content. Part 2 incorporated evaluation of YouTube trending videos that were associated with science and educational content for behavioral, emotional, and cognitive engagement using quantitative analyses.

Data collection and sample

Study part 1

Aiming to estimate the proportion of science content on YouTube, we created a usable scope of data that focused on trending videos. Trending videos were selected for our data set as it is the only approach that is allowed by the YouTube Data Application Programming Interface (API) for a systematic collection of public standard feeds. Although this is a convenience sample, it is pertinent and informative, because it yields the videos which have the potential to become high traffic videos on YouTube in a relatively short time. YouTube Data API permits retrieval of only 200 new top trending videos a day, including feeds and statistics related to videos and users. To characterize the collected video content, we used a YouTube Data API feature to automatically associate video categories . This enabled us to learn about the landscape of daily spontaneous topics that people in different viewership communities are watching on YouTube, for example, views of a how-to topic that is relevant to everyday decisions, such as troubleshooting and problem-solving; or views of an educational topic that might be used for recording and disseminating course lectures (Kousha et al., 2012). Our data set was collected from March 2019 to July 2019 using a Python script (Jolly, 2018).

Study part 2

To assess the level of behavioral engagement with science content, we counted the number of views, likes, dislikes, number of comments for each trending video that was classified by YouTube’s API as Science and Education content. The data were collected at two different time points: the first time point was during part 1 of the study, as described earlier; the second time point was 1 year later, during May 2020. To assess the level of emotional engagement and cognitive engagement, we focused on a random sample of 1000 comments from each video that was associated with science and educational content. Each comment was double-coded for emotional engagement using an automated sentiment analysis (see below), and for cognitive engagement using an automated classification of argumentative moves. An automated sentiment analysis to pinpoint emotional engagement was conducted using natural language processing with the Syuzhet R package (Jockers, 2017). The Syuzhet package incorporates several lexicons which enables both sentiment polarity evaluation (i.e. reporting positive or negative words), and emotional categorization into the following eight types: anger, anticipation, disgust, fear, joy, sadness, surprise, trust, positive, negative. Syuzhet natural language processing has been applied to several sources of data and was found reliable (Ragupathy and Maguluri, 2018; Widyaningrum et al., 2019; Yoon et al., 2017). Grounded in the Interactive-Constructive-Active-Passive (ICAP) framework that suggests that arguing a position is the highest level of cognitive engagement (Chi et al., 2018), we tracked elementary units of argumentative discourse. We developed an automated coding scheme of linguistic argumentation elementary units (i.e. agree, disagree, evidence, reason, fact (Mochales and Moens, 2011; Toulmin, 1958; Weinberger and Fischer, 2006). Then, we used nCoder, an online software (Arastoopour Irgens et al., 2020; Cai et al., 2019) to automatically apply the coding scheme to 89,000 comments. Inter-rater reliability between a human rater (first author) and the automated algorithm was almost perfect, Cohen’s kappa = 0.96 (McHugh, 2012).

Analysis

Descriptive statistics were computed using means, standard deviations, frequencies, and percentages to describe the distribution of the top trending YouTube videos. In addition, a one-way analysis of variance (ANOVA) and repeated measures ANOVA were conducted to examine differences across different science and educational topics comparing two time points of data collection. To evaluate associations between engagement characteristics, we used a bivariate parametric correlation analysis. Data were analyzed using SPSS version 25.

3. Findings

Study part 1

As we mentioned in the Data Collection section, we collected our trending video data over a 3-month period starting from March 2019; and each day, we added a new collection of 200 trending videos to our data set. Thus, in total, we collected 14,700 trending videos (for details see Supplemental material 1). Analysis revealed that videos with educational and science content were included among the top trending YouTube videos. We learned that the three leading categories (of the categories automatically generated by the API) that incorporated the highest number of videos were (1) the entertainment category (encompassing 30% of the top trending videos); (2) the music category (encompassing 14.5% of the top trending videos); (3) and the sports category (encompassing 11.4% of the top trending videos); which derived from 338, 248, and 148 channels, respectively (see Supplemental material 1). Science and education was the 11th highest category deriving from 36 channels. Videos with science and educational content derived from only 36 channels and encompass 3.4% (N = 503) of the top trending videos (Figure 1). Even though the science and education category has a relatively small amount of trending videos, the analyses show that they were able to trigger a relatively high number of views on average, reaching the sixth highest category (more than the comedy and sports categories, Figure 1a) and also a high number of post-video comments on average, reaching the eighth highest category (Figure 1b).
Figure 1.

(a) mean number of views per category and (b) Mean number of post-video comments per category.

(a) mean number of views per category and (b) Mean number of post-video comments per category.

Study part 2

Behavioral engagement

To pinpoint characteristics of behavioral engagement with science content, we focused on 503 videos that were classified as science and education content. Of these, we removed videos that were selected as trending videos multiple times (duplicates), which left us with 89 videos. Subsequently, for the sample of 89 videos, we collected the following data at two different time points, a year apart: number of comments, views, likes, dislikes (for details see Table 1).
Table 1.

Summary data of top trending videos with focus on science and educational content at two different time points (89 videos).

TopicNo of videos (%)No of commentsViewsLikesDislikesExample
Time 1Time 2Time 1Time 2Time 1Time 2Time 1Time 2
Human organism15 (17)11,095 (±16,382)19,171* (±26,028)1,696,285 (±1,444,401)4,718,483 (±4,725,446)87,394 (±104,025)176,016 (±214,982)2380 (±3031)4429 (±4468)How fall asleep/vaccines side effects/human hair
Universe16 (18)6771 (±6811)8010 (±7243)1,826,481 (±1,755,865)3,158,983 (±2,689,998)62,001 (±65,661)89,231 (±93,564)1415 (±1745)2052 (±2015)The quantum theory that connects the entire universe
Nature10 (11)5678 (±5376)7867 (±7452)2,141,678 (±3,657,271)4,050,883 (±6,084,642)66,722 (±107,129)101,104 (±149,392)1767 (±1873)2738 (±2375)World’s lightest solid/ liquid oxygen & diamond?
Making & How it works30 (34)8414 (±11,085)10,353 (±14,041)3,109,563 (±5,472,657)5,510,661 (±8,332,356)82,701 (±115,964)123,449 (±161,183)2771 (±3778)4011 (±5250)Magnetic sand/magnesium fires
History9 (10)1647 (±667)1966 (±933)584,937 (±277,317)1,436,425 (±662,186)16,055 (±8208)26,259 (±8559)485 (±229)485 (±314)U.S. Civil War
Other9 (10)3223 (±2234)4320 (±3316)1,704,515 (±2,767,927)2,989,407 (±4,778,036)29,563 (±19,221)46,810 (±35,609)1916 (±3702)2991 (±5804)Language and accent/economics

Numbers represent Mean ± SD.

Significant difference p < .05 between time points compared to differences in other video categories.

Summary data of top trending videos with focus on science and educational content at two different time points (89 videos). Numbers represent Mean ± SD. Significant difference p < .05 between time points compared to differences in other video categories. YouTube offers various interactions with the content, which might be distinguished from simple exposure to content, such as viewing the video, and by more proactive behaviors of commenting, liking and disliking. Following one-way ANOVA that compared the total number, within our full corpus of data, of each of these types of interactions, Bonferroni post hoc tests showed that viewing videos was the most common user experience compared to liking, commenting, and disliking (2,134,570 ± 3,665,698; 65,862 ± 93,105; 7054 ± 10,133; 2031 ± 2970, respectively; F(3,352) = 29.5, p < .0001). Using one-way ANOVA, we compared counts of instances of behavioral engagement in post-video comments between videos covering different topics: human organism, universe, nature, making, history and others (see Supplemental material 2). There were no significant differences between topics in the number of instances of cognitive engagement in post-video comments (F(15,332) = 0.81, p = .66). We further conducted repeated measures ANOVA to test whether there were differences, within each topic, in the increase between time points in the number of views, comments, likes, and dislikes. The only significant increase between time periods was the increase in comments on the topic human organism (health) (F(5,83) = 2.42, p < .05; for details see Supplemental material 3). All other differences between topics in increases between time points in views, comments, likes, and dislikes, were not significant. We tested for correlations between views, comments likes and dislikes within the science and educational videos using a bivariate parametric correlation analysis. There were moderate to strong correlations between the various engagement behaviors (Table 2). Namely, there is a moderate to strong correlation between the number of comments posted to the number of views (r = .64, p < .01), number of likes (r = .88, p < .01), and dislikes (r = .75, p < .01; Table 2).
Table 2.

Bivariate intercorrelations between behavioral, emotional and cognitive engagement (N = 89 videos).

Variable12345678910111213141516
Behavioral engagement1. Comments
2. Views0.64***
3. Likes0.88***0.71***
4. Dislikes0.75***0.84***0.76***
Emotional engagement1.5. Neutral emotions-0.17−0.01−0.16−0.08
6. Positive emotions0.13−0.090.110.01−0.86***
7. Negative emotions0.20−0.020.080.08−0.85***0.75***
8. Anger0.07−0.09−0.040.04−0.69***0.59***0.88***
9. Disgust0.24*−0.040.100.08−0.68***0.67***0.83***0.78***
10. Fear0.17−0.090.040.05−0.71***0.65***0.89***0.87***0.76***
11. Sadness0.24*−0.020.130.08−0.77***0.73***0.91****0.76***0.80***0.79***
12. Anticipation0.10−0.090.07−0.03−0.77***0.89***0.74****0.60***0.60***0.65***0.74***
13. Joy0.21*−0.050.170.05−0.80***0.87***0.66***0.57***0.60***0.60***0.62***0.81***
14. Surprise0.10−0.030.080.05−0.79***0.82***0.81***0.66***0.67***0.72***0.78***0.86***0.74***
15. Trust0.15−0.080.110.02−0.83***0.95***0.80***0.66***0.70***0.71***0.77***0.91***0.85***0.82***
Cognitive engagement16. Argumentation0.22*−0.050.29**0.04−0.71***0.70***0.78***0.46***0.50***0.55***0.70***0.74***0.55***0.66***0.81***

Data represents the Pearson r.

p < .05; **p < .01; ***p < .001.

Bivariate intercorrelations between behavioral, emotional and cognitive engagement (N = 89 videos). Data represents the Pearson r. p < .05; **p < .01; ***p < .001.

Emotional engagement

Semantic analysis was conducted on a sample of 1000 comments from each video (89 videos, total 89000 comments). Sentiment polarity evaluation shows that on average there were significantly fewer negative expressions (407 ± 205) than positive and neutral expressions (580 ± 304; 545 ± 99, respectively) per video (F(2,258) = 15.16, p < .001). Furthermore, we compared how often post-video comments express each of eight types of emotions: anger, anticipation, disgust, fear, joy, sadness, surprise, and trust. This one way ANOVA analysis shows that there is a significant difference in how emotions are expressed in post-video comments (F(7,688) = 30.1, p < .001). Bonferroni post hoc analysis found that overall trust is the most commonly expressed emotion, that there are significantly more comments that express joy than anger, and that anger and disgust are half as often expressed than trust and anticipation (Figure 2). There were no significant differences between science and educational topics in the level of emotional engagement expressed within their associated post-video comments (one way ANOVA F(35,648) = 1.22, p = .18).
Figure 2.

Sentiment analysis: Amount of emotional expressions per video (Mean ± SD; median; 10–90 percentile range).

Sentiment analysis: Amount of emotional expressions per video (Mean ± SD; median; 10–90 percentile range).

Emotional engagement and behavioral engagement

Bivariate parametric correlation analysis demonstrates significant small to moderate correlations between number of posted comments and expressed feelings of sadness (r = 0.24, p < .05), disgust (r = 0.24, p < .05), and joy (r = 0.21, p < .05). That is, expression of emotions such as sadness, disgust, and joy were correlated with a higher tendency to post comments (Table 2). However, when analyzing increases in views, comments, likes, and dislikes between the two data collection time points, no significant correlations were found (Table 2).

Cognitive engagement

Analysis of cognitive engagement was conducted on the same sample as the semantic analysis of emotional engagement (89 videos, total 89,000 comments). Automated coding of linguistic argumentation using Syuzhet R package (Jockers, 2017), detected that at least 16% of comments included an argumentative expression. The one-way ANOVA analysis comparing the number of argumentative expressions in post-video comments between videos of different science and education topics shows differences with a large effect size (F(5,81) = 5.97, p < .001, η2 = 0.27). Bonferroni post hoc analysis found that videos with making content incorporated significantly less argumentative expressions than all other topics; making M = 39.6 (±28.3); universe M = 117.9 (±71.9); human organism, M = 84.1 (±71.3); nature M = 58.3 (±34.1); history M = 57.8 (±45.8); other M = 44.7 (±24.4).

Cognitive, emotional, and behavioral engagement

Bivariate parametric correlation analysis demonstrates significant positive correlations between cognitive, behavioral, and emotional engagement. The findings demonstrate a small but significant correlation between argumentation and behavioral engagement, namely: number of comments are correlated with average number of argumentative expressions (r = 0.22, p < .05); and number of likes are correlated with average number of argumentative expressions (r = 0.29, p < .01). However, when analyzing the amount of gained views, comments, likes, and dislikes between the two time points of data collection to cognitive engagement, no significant correlations were found (r = −0.12, p = .262; r = 0.05, p = .660; r = −0.05, p = .646; r = −0.07, p = .535, respectively). In addition, incorporation of argumentative expressions is correlated with positive emotional expressions (r = 0.84, p < .001) as well as with negative emotional expressions (r = 0.77, p < .001). Neutral statements were negatively correlated with argumentation (r = −0.82, p < .001). Detailed emotional analysis specified that expression of all eight emotions (anger, disgust, fear, sadness, anticipation, joy, surprise and trust) is strongly correlated with argumentative discourse (Table 2). To further understand the impact of emotional and behavioral engagement on cognitive engagement, we performed a hierarchical multiple regression analysis (Table 3). Preliminary analyses were conducted to ensure that there were no violations of the assumptions of normality, linearity, and homoscedasticity. In addition, before conducting the regression analyses, the variance inflation factor (VIF) value was measured to identify the presence of multicollinearity (Thompson et al., 2017). Multicollinearity was an issue for emotional variables; thus, the anticipation variable was removed to make sure that the tolerance was greater than 0.10 for all variables in the model, and VIFs were well below 6.0 (Stevens, 2001). In Model 1, characteristics of behavioral engagement did not have a significant contribution to the regression model, F(4, 82) = 1.86, p = .12. However, incorporation of emotional engagement characteristics in Model 2 explained 70% of the variance in cognitive engagement. Trust, joy, sadness, and disgust had a positive effect on cognitive engagement. Thus, the model as a whole explained 79% of cognitive engagement level, F(7, 75) = 35.47, p < .001.
Table 3.

Summary of hierarchical regression analysis of cognitive engagement (N = 89 videos).

VariableModel 1Model 2
β t β t
Comments0.351.480.251.83
Views0.291.460.060.59
Likes0.010.06−0.12−0.95
Dislikes0.010.06−0.03−0.01
Anger0.181.43
Disgust0.222.17*
Fear0.010.08
Sadness0.322.71**
Joy0.504.59***
Surprise0.080.64
Trust1.178.88***
AIC702.449591.342
R 2 0.080.78
F for change in R21.8635.47***
ΔR20.080.70

AIC: Akaike Information Criterion.

Coefficients in the table are standardized beta coefficients.

p < .05; **p < .01; ***p < .001.

Summary of hierarchical regression analysis of cognitive engagement (N = 89 videos). AIC: Akaike Information Criterion. Coefficients in the table are standardized beta coefficients. p < .05; **p < .01; ***p < .001.

4. Discussion

The goal of this study was to extend our understanding of public engagement with science on social media. In particular, we focused on the characteristics of social media, which enable person-to-person engagement, and on analyses of actual media interactions rather than self-report data on surveys. Self-reports offer important insights, especially into the meaning the individuals ascribe to their interactions, but analyses of actual media interactions offer a way to more accurately assess sequence, frequency, and forms of interaction, as well as the ability to test their association with multiple factors. We explored the magnitude of engagement with science relative to other content on YouTube, and then, we investigated in-depth characteristics of behavioral, emotional, and cognitive dimensions of engagement. Overall, our findings show positive correlations between the three types of engagement, and illuminate the role of emotions in public engagement with science on social media. Here, we elaborate on these findings. To explore the magnitude of interactions with science on YouTube, we analyzed both the distribution of different types of trending videos, and the proportion of content that might be relevant to engaging with science. Our analysis revealed that the most trending content that users search for on YouTube relates to entertainment and music. However, the top trending YouTube videos (i.e. 3.4% of top trending videos) also include videos on educational and science topics, which attract a relatively high number of views and post-video comments. This finding is consistent with the Science and Engineering Indicators report about Internet and social media as a primary source of information about science and technology (National Science Board, 2018). Such information seeking motives are among the most likely to predict participatory acts such as liking or disliking videos, and commenting on videos (Khan, 2017). Investigating actual interactions, Thelwall et al. (2012) corroborate the interest in science and the propensity for participatory acts. They show that the most discussed topics, which received the highest number of replies on YouTube, were related to the science and technology category. Our findings extend this earlier research by identifying the expression of and interrelationships between behavioral, emotional, and cognitive engagement. First, we explored behavioral engagement with science content. We found that on average, each video with science content received 20 times more views than likes, and 300 times more views than comments. Indeed, studies show that consumption behavior, absent participatory acts such as likes or comments, makes up 90% of many online communities (Edelmann, 2016; Nonnecke and Preece, 1999). Of course, we do not know whether this “viewing only” participation includes additional offline behaviors that would qualify as deeper engagement. It would be interesting to pursue future studies that investigate both online and offline behavior in order to gauge whether social media participatory acts reflect the full extent of engagement with the content posted on social media. Past research, based on self-reports of information elaboration provide some insights (Jensen, 2011; Oeldorf-Hirsch, 2018; Perse, 1990), but direct observation and measurement of behavior would provide more veridical accounts. Activity that is visible on social media, posting comments and participating in post-video discussions, reflects the rarer type of engagement, cognitive engagement. Our findings show that behavioral and cognitive engagement are related. Although correlational analyses cannot reveal causal or temporal directions, we believe that the greater prevalence of behavioral engagement suggests that behavioral engagement triggers cognitive engagement. In other words, viewers of science videos are more likely to engage in person-to-person argumentative deliberation if the video content has prompted behavioral engagement such as posting comments and liking the video. Cognitive engagement, which may be most pertinent to public engagement with science, is also associated with emotional engagement. Consistent with earlier research analyzing user comments (Knowles and Wilkinson, 2015; Orr and Baram-Tsabari, 2018), our findings show that YouTube post-video comments offer an important space for emotional expressions, and that these are related to other forms of engagement with science on social media (Maier et al., 2014). Our findings revealed that expression of both positive and negative emotions, as in the case of behavioral engagement, correlate with commenting behaviors. Negative emotions, such as disgust or sadness, and positive emotions, such as joy, were linked to a tendency to post more comments and, as such, to deeper cognitive involvement. Thus, emotional stimulation, regardless of valence, is associated with, and possibly triggers both behavioral and cognitive engagement. This idea is further supported by the finding that neutral emotional expressions were negatively related to processes of argumentative elaboration. More research is needed before we can make specific recommendations for science communication. However, some immediate implications from this study are that in composing videos, science communicators should prioritize emotional triggers alongside considerations concerning the content of the message, such as detail, accuracy, and clarity. The study further implies that science communicators have some flexibility in terms of the specific emotions that they target, because both positive and negative emotions were associated with engagement. This study has several limitations that should be taken into consideration. First, YouTube API does not make users’ demographic data available; therefore, future studies should look into variables such as gender and age to evaluate their effect on engagement level. For example, do gender gaps identified among YouTube video hosts (Amarasekara and Grant, 2018) also appear in differences in forms or degree of engagement. In addition, the analysis did not capture the accuracy and quality of the comments; therefore, further studies should consider whether the content of the comments is accurate or not, and evaluate the effect of comment quality on engagement. This line of research can also be extended by incorporating qualitative analyses, and by comparing the different types of user engagement across social media platforms such as between Facebook and YouTube, in trying to understand the relationship between platform characteristics and forms of engagement. Furthermore, this study focuses on the role of emotional valence and emotional presence only; however, it has been shown that emotional strength has different effects on citizen engagement with social media (Ji et al., 2019). Thus, subsequent studies should analyze the role of emotional strength. In addition to addressing these limitations, there are also other future lines of research suggested by the current findings. One line of research could collect social media data over smaller intervals of time in order to examine interrelationships between changes in science content and changes in engagement. Our study highlights the role of emotion in social media interactions, and its interactions with other forms of engagement. Therefore, studies that explore the effects of emotional engagement on other factors would extend our understanding of the role of social media in public involvement with science. For example, Huber et al. (2019) suggest that emotion could be a key variable in understanding dynamics of trust, and Reif et al. (2020) relate emotions to perceptions of scientific trustworthiness (Reif et al., 2020).

5. Conclusion

The ubiquity of social media and calls for increasing public engagement with science motivate this research. We aimed to clarify the construct of engagement, and investigate how behavioral, emotional, and cognitive engagement with science on social media are distinguished, but interrelated. Cognitive engagement on social media hinges on behavioral and emotional engagement. This relationship holds regardless of valence. Both positive (e.g. joy) and negative emotions (e.g. sadness) are associated with and may trigger greater behavioral and cognitive engagement. Therefore, social media interactions, which tend to evoke emotional response, offer promise in advancing person-to-person cognitive engagement with science. Click here for additional data file. Supplemental material, sj-pdf-1-pus-10.1177_0963662521990848 for Interactions between emotional and cognitive engagement with science on YouTube by Ilana Dubovi and Iris Tabak in Public Understanding of Science
  17 in total

1.  Organ donation on Web 2.0: content and audience analysis of organ donation videos on YouTube.

Authors:  Yan Tian
Journal:  Health Commun       Date:  2010-04

2.  Science and society. Framing Science.

Authors:  Matthew C Nisbet; Chris Mooney
Journal:  Science       Date:  2007-04-06       Impact factor: 47.728

3.  Headlice eradication as everyday engagement with science: An analysis of online parenting discussions.

Authors:  Christine Hine
Journal:  Public Underst Sci       Date:  2014-07

4.  Medical Drama Viewing and Medical Trust: A Moderated Mediation Approach.

Authors:  Yan Tian; Jina H Yoo
Journal:  Health Commun       Date:  2018-10-25

5.  Translating the ICAP Theory of Cognitive Engagement Into Practice.

Authors:  Michelene T H Chi; Joshua Adams; Emily B Bogusch; Christiana Bruchok; Seokmin Kang; Matthew Lancaster; Roy Levy; Na Li; Katherine L McEldoon; Glenda S Stump; Ruth Wylie; Dongchen Xu; David L Yaghmourian
Journal:  Cogn Sci       Date:  2018-06-28

6.  Fostering public trust in science: The role of social media.

Authors:  Brigitte Huber; Matthew Barnidge; Homero Gil de Zúñiga; James Liu
Journal:  Public Underst Sci       Date:  2019-10

7.  Classifying Academically At-Risk First Graders into Engagement Types: Association with Long-Term Achievement Trajectories.

Authors:  Wen Luo; Jan N Hughes; Jeffrey Liew; Oiman Kwok
Journal:  Elem Sch J       Date:  2009-03-01

8.  What drives science media use? Predictors of media use for information about science and research in digital information environments.

Authors:  Julia Metag
Journal:  Public Underst Sci       Date:  2020-07-01

9.  Engagement with health agencies on twitter.

Authors:  Sanmitra Bhattacharya; Padmini Srinivasan; Phil Polgreen
Journal:  PLoS One       Date:  2014-11-07       Impact factor: 3.240

10.  Interrater reliability: the kappa statistic.

Authors:  Mary L McHugh
Journal:  Biochem Med (Zagreb)       Date:  2012       Impact factor: 2.313

View more
  3 in total

1.  Guidance in the chaos: Effects of science communication by virologists during the COVID-19 crisis in Germany and the role of parasocial phenomena.

Authors:  Sonja Utz; Franziska Gaiser; Lara N Wolfers
Journal:  Public Underst Sci       Date:  2022-05-18

2.  Online learning satisfaction and its associated factors among international students in China.

Authors:  Mei Tian; Genshu Lu
Journal:  Front Psychol       Date:  2022-07-22

3.  How Portuguese Health Entities Used Social Media to Face the Public Health Emergency during COVID-19 Disease.

Authors:  Daniela Azevedo; Ana Isabel Plácido; Maria Teresa Herdeiro; Fátima Roque; Vítor Roque
Journal:  Int J Environ Res Public Health       Date:  2022-09-21       Impact factor: 4.614

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.