Literature DB >> 36118938

Studying topic engagement and synergy among candidates for 2020 US Elections.

Manmeet Kaur Baxi1, Rajesh Sharma2, Vijay Mago1.   

Abstract

This article provides a comprehensive summary of how candidates running in the 2020 US Presidential Elections used Twitter to communicate with the public. More specifically, it aims to uncover elements linked to public engagement and internal cooperation (in terms of content and stance similarity among the candidates from the same political front, and with respect to the official Twitter accounts of their political parties). Our main subjects are the Presidential and Vice-Presidential candidates who contested for the 2020 US Elections from the two major political fronts-Republicans and Democrats. Their tweets were evaluated for social reach, content similarity and stance similarity on 22 topics. According to the findings, Joe Biden had the highest engagement and impact (user impact: 177.08k, normalized to 0.99), followed by Donald Trump (user impact: 164.19k, normalized to 0.92). The Democrats depicted a clearer understanding of their audience, portraying an essential link between public participation, internal cooperation and the electoral campaign. The results also demonstrate that specific topics (like US Elections, and Inauguration Ceremony) were more engaging than others (Trump Healthcare Plan, and The Supreme Court Appointments). This study adds to the existing work on using social media platforms for electoral campaigns and can be effectively utilized by contesting candidates.
© The Author(s), under exclusive licence to Springer-Verlag GmbH Austria, part of Springer Nature 2022, Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Entities:  

Keywords:  2020 US elections; Content similarity; Electoral campaigns; Public engagement; Social media

Year:  2022        PMID: 36118938      PMCID: PMC9464427          DOI: 10.1007/s13278-022-00959-9

Source DB:  PubMed          Journal:  Soc Netw Anal Min


Introduction

Social media platforms (SMPs) like Facebook, Twitter and Instagram have become the conventional modes of online campaigns for elections after being first used by Barack Obama while contesting for his 2008 candidacy (Effing et al 2011). Researchers from around the world have expressed a strong interest in analyzing and evaluating social media data in the context of elections across these different platforms, as discussed in the works of Borah (2016); Russmann and Svensson (2017), and Vesnic-Alujevic and Van Bauwel (2014). Various politicians have widely used SMPs to express their views on current topics, share the latest developments in their constituencies, and communicate with their potential voters strategically. From the statistics in the report by Pew Research Center (2020), out of the top 10% adult Twitter users in the US, 92% of them are politicians, with Democrats or Democratic-leaning independents being hyperactive and capturing 69% positions in the top Twitter users; while Republicans or Republic-leaning independents occupying the remaining 26%. Additionally, recent studies have shown that maintaining an active presence on SMPs has helped politicians address social concerns and build a stronger relationship with the audience (Bonsón et al. 2019; Gruzd et al. 2018; Sahly,Shao and Kwon 2019). Furthermore, the cooperation among the candidates from the same political front has helped them communicate their policy initiatives clearly and organize support from the related interest groups (Grossmann 2014). The authors of (Wonka and Haunss 2020) emphasize the benefits of cooperation among the political parties and interest groups in European Union policy-making by examining their information networks. Hence, there is a requirement to investigate the influence of citizens’ engagement and the internal cooperation among the politicians on the election results. The authors of Bonsón et al (2012); Norris and Reddick (2013) have stressed the importance of a future qualitative study to quantify the genuine impact of social media on Government to Citizen (G2C) interactions. As a result, the goal of this research is to measure the utility of Twitter for politicians during the 2020 Presidential elections in the United States. In particular, we investigate the following two research questions:

Research Question 1 —What topics did the candidates discuss through online (Twitter) and offline (Presidential debates) mediums? How engaging were these topics and to what extent during the different phases of the electoral campaign?

Understanding what information and topics appeal to the audience the most, may be an effective method for gaining attention and increasing involvement (Bonsón et al 2015). Bonsón and Bednárová (2018) found that certain types of contents are more engaging than others. Therefore, identifying such materials and developing thorough plans for maintaining a continuous dialog with the citizens, responding to their grievances, recommendations and wants, would help improve governance quality. Following the identification of objectives and goals, norms could be produced to facilitate the contesting candidates with an effective tool for communication on SMPs. If implemented properly, it is indeed a win–win situation for both the candidates and the citizens. On that account, we calculate the impact of candidates and the engagement received by them on various topics, followed by classifying them according to the topic stickiness in different phases of the election campaign.

Findings

Joe Biden was the most impactful candidate among all, and Democrats tweeted more about topics of public interest during the electoral campaign as compared to Republicans. The detailed observations can be found in Sect. 4.

Research Question 2—Did the candidates from the same political front have similarities in their tweets and the stance for the topics with respect to their political front ?

Politicians collaborate to share resources and coordinate political support. The various types of cooperation networks formed inside a political front during the European Union policy-making have been highlighted by Wonka and Haunss (2020). Furthermore, the smaller networks inside a political party or interest group may reconfigure themselves based on the reputation (impact of the candidates) and the internal reciprocity (similarity in thoughts/actions) (Balliet et al. 2014; Gallo and Yan 2015), and Gross and De Dreu (2019). Thus, to analyze the synergy among the candidates during the electoral campaign of 2020 US Elections, we employ two methods, i.e., content similarity-based on the tweets, and stance similarity-the standpoint of candidates with respect to different topics. Understanding these aspects would help us identify which political front was more cooperative among themselves and echoed similar thoughts on Twitter. Kamala Harris depicted a higher amount of cooperation with both Joe Biden and the official account of Democrats in both — content and stance. On the other hand, Republicans portrayed comparatively lower synergy in their stance with respect to different topics. Refer to Sect. 5 for more details. Overall research framework This research provides a comprehensive analysis of the contesting candidates, a combination of both qualitative and quantitative insights into their online behavior. Previously, this form of hybrid research has proven to be beneficial (Sampieri 2018). Our study uses statistical methods to examine the public metrics of candidates’ tweets (the number of likes, replies, retweets, and quotes) and determine their social reach. It is qualitative as we infer the topics the candidates tweeted about and the similarity in the content and stance of the candidates from the same political front. Therefore, the objectives of this study are to discover the topics discussed by the candidates through online (Twitter) and offline (Presidential debates) channels, as well as the civic engagement on these topics during the different phases of the electoral campaign, and throughout the whole election campaign. Additionally, we also investigate the similarities in the tweets of the candidates from the same political front with respect to the tweet content and their stance on the topics. Furthermore, we try to uncover any relationships between public engagement and internal cooperation (content and stance similarities) that might have aided the candidates in contesting the 2020 US Presidential elections. Figure 1 presents an overview of the research framework.
Fig. 1

Overall research framework

The remaining part of the article is arranged as follows: Section 2 gives a summary of prior work linked to the use of SMPs in the electoral campaigns. Section 3 explains the data utilized in this study. The techniques used and insights of public engagement during the electoral campaign are presented in Sect. 4. Section 5 discusses the approach followed and observations for the identifying the collaboration among the candidates from the same political front. Finally, the key inferences and further research directions are presented in Sect. 6.

Related work

The analysis on elections has been widespread across different SMPs, like Bruns and Moon (2018) compared the candidate and audience activity on Twitter for the 2013 and 2016 federal elections in Australia. Additionally, the role of SMPs in enhancing democratic participation during the 2012 and 2016 Ghana elections was studied by Dzisah (2018). Praznik et al (2021) took a different approach by analyzing the strength of networks based on the usage of hashtags on Twitter during electoral campaigns (Praznik et al 2021). Gaisbauer et al. (2021) tried to contrast different opinion groups using network representations for replies and retweets in the context of Saxon state elections and violent riots in the city of Leipzig, Germany, in 2019, and Bilal et al (2019)’s work surveyed the current state-of-the-art approaches to analyze the election prediction mechanisms in use (Bilal et al 2019). As far as SMPs are concerned, Borah (2016) investigated the use of Facebook for campaign strategies used in 2008 and 2012 US Presidential elections, Russmann and Svensson (2017) analyzed Instagram for Swedish elections, and Vesnic-Alujevic and Van Bauwel (2014) studied the use of YouTube as an advertising tool during the campaign of European Parliament elections. According to previous studies (Bertot et al. 2010; Bonsón et al 2012; Chun et al 2010), SMPs can help enhance the transparency, involvement, and correspondence in governance. Also, researchers have revealed the influence on different features of public interaction through various instances (Gruzd and Roy 2016; Hollebeek et al. 2014; Ríos et al. 2017). Several authors (Bonsón et al. 2019; Bonsón and Ratkai 2013; Bonsón et al. 2017; Gruzd et al. 2018; Sahly,Shao and Kwon 2019; Siebers et al. 2019) have highlighted the relevance of SMPs as a vital instrument for amplifying social reach, and to help understand the audience better. However, earlier studies have also found that the sentiment, emotion, stance of the tweets, the promotion of tweets by bots; and collaboration between interest groups (polarization) may mitigate the impact of different aspects of public involvement and steer the change in public opinions (Bhat et al 2021; Galgoczy et al. 2022; Grover et al. 2019; Sandhu et al. 2019; Sandoval-Almazan and Valle-Cruz 2018). Internal cooperation and collaboration play an essential role for a political party to convey their policy initiatives during the electoral campaigns and gather support. The survey by Khanam et al. explores various methodologies proposed till date on the usage of the Homophily principle (likelihood of similar-minded people to engage with one another in communities) across different domains (Khanam et al 2022). Another survey by Chandrasekaran and Mago (2021) lists the different methods available to evaluate the semantic similarity between texts, ranging from traditional Natural Language Processing (NLP) techniques to deep neural-network-based hybrid methods (Chandrasekaran and Mago 2021), out of which, we alter one method as per the objectives of this study, to measure the internal cooperation among candidates based on their tweets. However, several factors affect the conflicts and synergy within a group as highlighted in the studies (Gross and De Dreu 2019; Larson 2021; Madeo and Mocenni 2020; Perkoski 2019). It is important to identify them for effective operations and governance. Our study evaluates the cooperation among the candidates by comparing the content of their tweets and collating the similarities in their stance on different topics. We also uncover the relationship between internal synergy and electoral campaigns. Formerly, Bansal and Srivastava (2018); Joseph (2019); Nugroho (2021), and Tsai et al (2019) have performed sentiment analysis on tweets using machine learning techniques, like lexicon-based models-VADER, and decision trees to predict the election results by focusing on a single aspect. Additionally, Chen et al (2021) have released a dataset for analyzing the 2020 US Elections. However, there has not been an empirical analysis to understand and examine the utility of Twitter (with emphasis on public participation and internal cooperation) as a communication tool during the 2020 US electoral campaign, considering different factors like social reach and internal cooperation (stance and content similarity), which is hence the focus of this work.

Dataset

We collected a total of 117,217 tweets authored from the accounts of Presidential and Vice-Presidential candidates of two major political fronts—Republicans and Democrats using the Twitter API v21, during the time frame of January 21, 2019, to January 27, 2021. Additionally, the tweets created by the official Twitter handles of both the political fronts were also collected. The candidates selected for our analysis and the number of tweets scraped from their accounts are discussed in Table 1. For the scope of our research, the political fronts (Democrats/Republicans) confine to the Presidential candidate, Vice-Presidential candidate, and the official Twitter handles of the political fronts. The official Twitter handles of the political fronts are referred to as ‘OfficialDemocrats’ and ‘OfficialRepublicans’ for Democrats and Republicans, respectively.
Table 1

Tweet distribution of the candidates selected from both the political fronts

Political partyCandidates (Twitter handle)Number of tweets
DemocratsPresidential Candidate: Joe Biden (@JoeBiden)5,486
Vice-Presidential Candidate: Kamala Harris (@KamalaHarris)5,835
OfficialDemocrats (@TheDemocrats, @HouseDemocrats, @SenateDems)41,728
Total53,049
RepublicansPresidential Candidate: Donald J. Trump (@realDonaldTrump, @POTUS)21,007
Vice-Presidential Candidate: Mike Pence (@Mike_Pence, @VP, @VP45)12,003
OfficialRepublicans (@GOP, @HouseGOP, @SenateGOP)31,158
Total64,168

The italicized text signifies the rank of contestants for 2020 US Presidential Elections along with the total number of tweets for each political front

Tweet distribution of the candidates selected from both the political fronts The italicized text signifies the rank of contestants for 2020 US Presidential Elections along with the total number of tweets for each political front Model parameters for topic clustering with TF-IDF document embeddings

Engagement and stickiness of topics

Identifying topics

We analyze the most and least discussed topics by the candidates from the political fronts through two sources—offline and online as defined below: The offline source is the topics that were discussed in the Presidential debates by both the candidates as given by ‘The Commission on Presidential Debates’2 and the events synchronous with the US Elections (Current/Snapshot events). The Presidential debates hosted by The Commission on Presidential Debates offer equal opportunities for candidates to express their opinions on the pressing issues (or political agendas), whereas the TV/YouTube commercials and Zoom rallies could vary according to the availability of resources a candidate has access to, and The online source of topics is topic modeling on the tweets authored by the candidates. We first preprocess and then cluster the tweets using various clustering algorithms and leverage the topics yielded by the best-performing topic model. Topics selected for analysis Mean coherence scores and CPU time for different clustering algorithms with TF-IDF embeddings over five runs with varying random states Bold signifies the best performing Clustering Algorithm Preprocessing: Firstly, all the non-alphabets (numbers, punctuation, new-line characters and extra spaces) were removed from the text using the regular expression module (re 2.2.1). Then, the text was tokenized using nltk 3.2.5, followed by the removal of stopwords. Also, tiny words (i.e., words with a length of fewer than three characters) were removed from the text. This was followed by stemming the text using PorterStemmer and lemmatizing it using the WordNetLemmatizer from nltk. Topic Modeling: Researchers have relied on Term Frequency-Inverse Document Frequency (TF-IDF) for generating document embeddings for short-text (Lilleberg et al 2015; Sari and 2016). Tweets are categorized as short texts3, and after preprocessing them, we generate document embeddings using TF-IDF and then pass them to five different clustering algorithms, namely Latent Dirichlet Allocation (LDA), Parallel LDA, Non-negative matrix factorization (NMF), Latent Semantic Indexing (LSI), and Hierarchical Dirichlet Process (HDP), to generate topic clusters. Due to the short and noisy nature of the data, we ran these models five times over the data with varying random seeds. We check the coherence scores of topic models based on words, the ‘c_umass’ (Newman et al 2010) and ‘c_v’ (Röder et al 2015) measure, to confirm the performance consistency over multiple runs and finally use the best model to extract the top five topics. We used Latent Dirichlet Allocation (LDA)4 and LDA multi-core5(Parallel LDA) provided by Gensim. Non-negative matrix factorization (NMF) model6 uses the online NMF proposed in Zhao and Tan (2016) for large corpora. Latent Semantic Indexing (LSI) model7 implements fast truncated SVD (Singular Value Decomposition). And, for HDP8, we use the improved online variational inference model proposed in Wang et al (2011). The details of parameters used for each of the models are listed in Table 2, and performance for each clustering algorithm in terms of their coherence scores (‘c_v’ and ‘c_umass’) and the amount of CPU time taken are mentioned in Table 4.
Table 2

Model parameters for topic clustering with TF-IDF document embeddings

Clustering AlgorithmEpochsChunk sizeWorkers (number of CPU cores)Evaluation Period (seconds)Alpha (A-priori belief on document-topic distribution)Eta (A-priori belief on topic-word distribution, also known as beta)Kappa (gradient descent step-size)Minimum normalizing probability
LDA2051000NA100.010.9NANA
Parallel LDA20510007100.010.9NANA
LSINA1000NANANANANANA
NMF2051000NA10NANA10
HDPNA1000NANA0.01NA1NA
Table 4

Mean coherence scores and CPU time for different clustering algorithms with TF-IDF embeddings over five runs with varying random states

Clustering Algorithmc_vc_umassCPU time (min:sec)
LDA0.70–2.2616852:52
Parallel LDA0.5921–2.4195512:12
NMF 0.773022 –1.61094 07:37
LSI0.585223–2.5935500:27
HDP0.640714–17.322301:38

Bold signifies the best performing Clustering Algorithm

From Table 4, we can see that NMF had the highest coherence scores (‘c_v’ and ‘c_umass’), followed by LDA and HDP. Hence, we selected the top five topics yielded by NMF to search across the first page of Google search results. The content from the first page of Google search results was then retrieved to make sense of the extracted topic keywords to suggest a good topic name. For example, for the set of keywords yielded by the topic model: [‘Paris’, ‘climate’, ‘green’, ‘change’, ‘science’, ‘reforms’, ‘environment’, ‘sustainable’, ‘urgency’], we did a Google search with these keywords and looked up for content and connections between them to deduce a suitable topic-phrase, i.e., ((Paris) Climate Agreement). Distribution of topics as per their abstract categories Hence, for each of the 22 selected topics (combined from both the sources, i.e., online and offline), we assign them an abstract category out of ‘Social Issues,’ ‘Healthcare,’ ‘Elections’ and ‘National Security.’ The abstract categories chosen serve as the foundation for political campaigns (Liu and Lei 2018), and they are the most significant categories to consider when trying to persuade the public to vote for a particular political front. We utilize all these topics to analyze the engagement, stickiness, and to predict the candidate’s stance. Table 3 presents the details of the selected topics as per their source and the abstract category they fall into. Each topic category consists of topics from at least two abstract categories. There are nine topics in ‘Social Issues,’ five in ‘Healthcare,’ and four in both ‘Elections’ and ‘National Security.’ The distribution of topics as per their abstract categories can be seen in Fig. 2. Furthermore, we analyze the most and least talked about topics and abstract categories for each candidate and political party based on engagement and stickiness.
Table 3

Topics selected for analysis

Topic sourceTopic categoryTopicAbstract category
OnlineModeled Topics (Topics generated from NMF)Legalization of Medical MarijuanaSocial Issues
Equality rights for LGBTQSocial Issues
Weapon BanSocial Issues
Build Back Express TourSocial Issues
Affordable Health Care ActHealthcare
OfflinePresidential Debate (1st)The EconomySocial Issues
The Supreme Court AppointmentsNational Security
COVID-19Healthcare
Race & Violence in our citiesSocial Issues
The Integrity of ElectionsElections
The Trump Biden RecordsElections
Trump Healthcare PlanHealthcare
Presidential Debate (2nd)Fighting COVID-19Healthcare
American Families & The EconomyHealthcare
Race in AmericaSocial Issues
Climate ChangeSocial Issues
National SecurityNational Security
LeadershipNational Security
Snapshot EventsBlack Lives MatterSocial Issues
Capitol Hill IncidentNational Security
US ElectionsElections
Inauguration CeremonyElections
Fig. 2

Distribution of topics as per their abstract categories

Engagement on topics

The amount of engagement received on a particular topic helps us quantify how popular the topic was among the general public. To quantify a topic’s engagement on Twitter, we first define each of the selected candidate’s (user’s) engagement on Twitter and then aggregate the tweets published by them, as well as their engagement as per the topic categories defined. The engagement for each user is defined as the product of average engagement per day and their impact. User engagement was formerly quantified in terms of community features (the number of communities a user is a member of), author features (number of followers/ following, author influence) and content features (the number of retweets, mentions, URLs, hashtags, keywords, comments, and sentiment subjectivity) (He et al. 2020; Purohit et al 2011). Similarly, we aim to include all the accessible features through the Twitter Academic API in this work, and the average engagement per day for a user () is computed as the product of average engagement for a tweet each day () and the user impact (userImpact). The average engagement for a tweet is aggregated by measuring the reactions received on tweets (such as the number of replies, retweets, likes, and quotes) over the course of the day. The data obtained from the Twitter API for the reactions to each tweet have been aggregated from January 21, 2019, to January 27, 2021. We propose average engagement per day for a tweet by taking inspiration from the Engagement rate defined by Twitter9. For a given user, Twitter defines Engagement Rate as:where Engagement is the summation over the number of likes, replies, retweets, media views, tweet expansion, profile/ hashtag/ URL clicks, and new followers gained for every tweet, and Impressions is the total number of times a tweet has been seen on Twitter, such as through a follower’s timeline, Twitter search, or as a result of someone liking your tweet. Due to limitations with the API, we only have access to the public metrics, i.e., number of likes, retweets, replies, and quotes. Therefore, to calculate the average engagement rate for a user per day, we use the function avgEngagementPerDay as proposed in Algorithm 1 (line number: 1). To normalize the fluctuating values of Average Engagement, we calculate its exponential moving average (EMA) with a window span of 20 days for every candidate10 and remove the outliers using z-score, followed by smoothening the average engagement per day to the 8th degree using Savitzky–Golay filter Every user has a different number of followers, following and they receive varied responses from the users on Twitter (which may or may not be their followers); hence, it is essential to consider their impact (popularity) on Twitter to calculate the number of users they reach through their tweets. Researchers have tried to analyze the impact of users by proposing heuristic and neural-network-based models (Daniluk et al 2021; Razis and Anagnostopoulos 2014; Son et al. 2020). We define the impact of a user () inspired from the previous work done in Razis and Anagnostopoulos (2014) and define it as a function of followers, following, the total number of tweets, and the profile age, as in Algorithm 1 (line number: 1), where followers is the total number of followers a user has, listedCount is the number of public lists a user is a part of, following is the number of people that the user follows, is the ratio of followers to following (FtF ratio) to check whether a user is an active user (with more followers, producing content) or a passive user (with more following, consuming content). To avoid outliers, we take log base 10 and add one to prevent the metric from being zero when the value of followers equals to following. tweetCount is the total number of tweets produced by the user for the scope of our analysis, and profileAge is the difference between the profile creation date reported by Twitter and January 27, 2021, i.e., the last day for our data collection, quantified as the number of days. Our algorithm overcomes the shortcomings of Razis and Anagnostopoulos (2014) by incorporating the listedCount factor and changing the placement of tweetCount. The tweetCount has been deemed inversely related to the user impact, because a user tweeting sporadically but obtaining high interaction is more significant than a person tweeting recurrently but receiving low engagement. The engagement for a user is the product of average engagement per day and the user’s impact. The engagement value helps us in quantifying the user’s social reach. Joe Biden had the highest impact, followed by Donald Trump, Mike Pence, Kamala Harris, OfficialDemocrats and OfficialRepublicans. We normalize the user impact between the range 0 and 1 to calculate the engagement on tweets for each topic, where 0 is the lowest user-impact, and 1 is the highest. For Joe Biden, the top three topics receiving the maximum engagement during the scope of our analysis were – The Integrity of Elections, Weapon Ban, and US Elections. As for Kamala Harris, they were US Elections, Fighting COVID-19 and The Integrity of Elections, and however, for the OfficialDemocrats, the most engaging topics differed from these, and they were The Supreme Court Appointments, Equality Rights for LGBTQ, and Fighting COVID-19. Overall, for Democrats, the top three most engaging topics were The Integrity of Elections, US Elections, and Weapon Ban. Electoral campaign timelines for Presidential and Vice-Presidential candidates. The timeline is divided as per the general election phases and the ranks each candidate was contesting for. The details of campaigning for each candidate have been taken from the news reports of the campaigns on CNBC and Politico Appearance of Loose topics in different election phases In the case of Donald Trump, the top three topics receiving the maximum engagement were US Elections, The Integrity of Elections, and Affordable Healthcare Act. For Mike Pence, they were Inauguration Ceremony, US Elections and Fighting COVID-19. However, for the OfficialRepublicans, as they had the lowest impact, the engagement received on the tweets was too low to be quantified. Overall, for Republicans, the top three most engaging topics were US Elections, Inauguration Ceremony and Affordable Healthcare Act.

Stickiness of topics

Stickiness helps us to identify the favorite topics for each candidate within the scope of our analysis. We quantify stickiness based on the repetitiveness of the topics spanning across different election phases. The Presidential candidate’s timeline is divided into three phases, and the Vice-Presidential candidate’s timeline is divided into four phases. See Fig. 3 for more detailed information about the timelines. The topic stickiness is checked for three candidates only, because of the unavailability of exact campaigning dates for Mike Pence. We segregate the topics into three classes based on stickiness:
Fig. 3

Electoral campaign timelines for Presidential and Vice-Presidential candidates. The timeline is divided as per the general election phases and the ranks each candidate was contesting for. The details of campaigning for each candidate have been taken from the news reports of the campaigns on CNBC and Politico

Very Sticky: Topics are Very Sticky when they have been tweeted about in every election phase, Sticky: If the topics were tweeted in election phases, then they were classified as Sticky. Here, n is the total number of election phases (i.e., 3 for Presidential, and 4 for Vice-Presidential candidates). Loose: If the topics were tweeted only once for Presidential candidates, and twice for Vice-Presidential candidates across all the election phases, then they were classified as Loose. Joe Biden and Kamala Harris tweeted mostly about Social Issues and Healthcare as we can see these categories dominate both the Very Sticky and Sticky levels from Fig. 4a and b. However, National Security topics were loose in nature. Donald Trump tweeted differently from the Democrats and tweeted the most about Elections and Healthcare, followed by Social Issues.
Fig. 4

Abstract categories of topics segregated according to stickiness levels for all three candidates (a) Joe Biden, (b) Kamala Harris, and (c) Donald Trump

Abstract categories of topics segregated according to stickiness levels for all three candidates (a) Joe Biden, (b) Kamala Harris, and (c) Donald Trump Doing a micro-analysis, we found that Joe Biden tweeted about twenty topics out of twenty-two, with fourteen of them being Very Sticky, five being Sticky, and one being Loose. Legalization of Medical Marijuana and Trump Healthcare Plan were the topics that Joe Biden had not tweeted about, even once. However, Kamala Harris and Donald Trump tweeted about 21 topics. Kamala Harris had eleven topics in the Very Sticky category, seven in Sticky and three in Loose. For Donald Trump, the distribution of topics per their stickiness levels was slightly different, with thirteen being Very Sticky, five being Sticky, and three being Loose. Kamala Harris did not tweet about Trump Healthcare Plan, and Donald Trump did not tweet about the Legalization of Medical Marijuana. Regardless of their political fronts, Joe Biden, Kamala Harris, and Donald Trump had nine Very Sticky topics in common (i.e., The Integrity of Elections, Affordable Healthcare Act, Equality Rights for LGBTQ, Weapon Ban, Inauguration Ceremony, US Elections, American Families & The Economy, COVID-19, Race & Violence in our cities). Joe Biden and Donald Trump stuck to The Integrity of Elections; however, Kamala Harris stuck to Affordable Healthcare Act. Comparing the candidates from the same political front, Joe Biden and Kamala Harris had eleven common topics in the Very Sticky category. For the Presidential candidates, in addition to the nine common topics, they also had Fighting COVID-19 and Capitol Hill Incident repeating in the Very Sticky category. Furthermore, the topics classified as Loose from all three candidates appeared in only one of the election phases. For example, in Joe Biden’s case, topic Supreme Court Appointments appeared only in the second election phase as per his timeline, and similar behavior can be seen for the other two candidates (refer Table 5 for details). Also, the Loose topics, i.e., Trump Healthcare Plan, Supreme Court Appointments, Build Back Express Tour, The Economy, and The Trump & Biden Records, are among the rarely tweeted topics, and corresponding behavior can be seen for the Very Sticky topics, i.e., the top three Very Sticky topics for each candidate are the most frequently tweeted and highly engaging topics.
Table 5

Appearance of Loose topics in different election phases

CandidateElection Phase 1Election Phase 2Election Phase 3Election Phase 4
Joe BidenNo loose topicsThe Supreme Court AppointmentsNo loose topicsNot applicable (NA)
Kamala HarrisThe Economy, The Trump & Biden RecordsNational SecurityNo loose topicsNo loose topics
Donald TrumpThe Economy, Trump Healthcare Plan, Build Back Express TourNo loose topicsNo loose topicsNot applicable (NA)

Synergy among candidates

To quantify the cooperation among the candidates from the same political front, we highlight the content similarity and the congruities and contrasts in the stance of various topics they tweeted about, as discussed below. Content similarity between candidates using different BERT-based embeddings

Content similarity

We check the alignment of the Presidential and Vice-Presidential candidates with the political party by comparing the similarity of their tweets as per our proposed Algorithm 2. When comparing the similarity of two users, there is a high probability that the topic of tweets from one user may be repeated by the second user a couple of days before or after the first user’s tweet during the election campaign. So, to address this, we compare each tweet of with all the tweets of and store the maximum similarity between the tweet text. We repeat this process for all the tweets, then average the results to determine how similar two candidates’ content is. The tweets of Vice-Presidential candidates are compared with both the Presidential candidate and the political party, and however, the tweets of the Presidential candidate are compared with the political party only. We compute the content similarity (cosine similarity) between the candidates by using the top-5 models from HuggingFace11 (grouped by the sentence-similarity task, sorted by the number of downloads) to generate text embeddings. Figure 5 elaborates the performance details for each of them.
Fig. 5

Content similarity between candidates using different BERT-based embeddings

From all the models tested, ‘bert-base-mean-nli-tokens’ performs the best in computing the content similarity between the candidates, followed by ‘paraphrase-multilingual-MiniLM-L12-v2.’ The common trend noticed while computing the content similarity is that the tweets by Kamala Harris are more aligned with the Presidential candidate than the political party for three out of five models; however, for Mike Pence, it’s the opposite, i.e., Mike Pence aligns more with the political party instead of the Presidential candidate. Also, the Presidential candidates have a high similarity rate with the political party. Therefore, the results portray coordination in the candidates’ tweets and the tweets from their political parties.

Stance similarity

Stance Similarity is an important technique to analyze textual data and is frequently used in NLP to analyze the standpoint of a person toward a topic or an event. We test different models that classify the candidates’ stance for the selected topics in three categories: favor, against, and neutral. Although there is no universal number for how many tweets should be sampled, for testing a model, we observe a range across studies from under 2,000 labeled tweets (Alomari at al 2017; Peisenieks and Skadiņš 2014; Şaşmaz and Tek 2021) to several thousand (Golubev and Loukachevitch 2020; Nabil et al 2015; Rustam et al. 2021; Zhang et al 2020). For this study, we sample 3,015 tweets evenly distributed across the timeframe of our data collection and from all the candidates. Three different annotators then labeled these tweets, and the stance category having the majority among the three annotators was chosen as the overall response. The annotators had no known prior political biases, and they annotated the tweets solely on the basis of the tweet content. The Fleiss’ Kappa statistical test was performed to determine the inter-annotator agreement in labeling, and the kappa score is ‘0.7’. We divide the data into an 80:20 ratio for training and testing multiple classification methods, and we annotate the dataset using standard procedures, as defined above. Stance classification performance on the testing set (i.e., 20% of the sampled dataset) using different algorithms We try various traditional and modern algorithms to estimate the performance for stance classification on the labeled tweets. We use TF-IDF and Hashing Vectorizer for the conventional algorithms to generate embeddings as inputs to support vector machine (SVM), linear SVM, and logistic regression to compute the performance. Synthetic Minority Oversampling Technique (SMOTE) is used to oversample the tweets’ vectorized features. However, we do not notice a rise in classification performance after oversampling. We also report the classification performances on modern algorithms, like the Deep Neural Network (DNN)-based classifier from the Spark NLP pipeline, which takes Universal Sentence Encodings of tweets as inputs12, BERT-base-uncased13, XLNet (base-cased14, large-cased15) and fine-tuned Facebook’s Zero-shot learning-based, bart-large-mnli16. From the traditional algorithms, TF-IDF combined with Logistic Regression, Hashing Vectorizer with SVM, and Logistic Regression perform equally well with their stance classification accuracies on the testing set of the sampled data as 73%. From the modern ones, Facebook’s ‘bart-large-mnli’ performs at par with 75% classification accuracy on the test set (Table 6). We then use Facebook’s bart-large-mnli to predict the stance on the remaining tweets.
Table 6

Stance classification performance on the testing set (i.e., 20% of the sampled dataset) using different algorithms

Algorithm usedOversampled (Yes/No)Classification performance
Hashing Vectorizer and Linear SVMNo0.70
Yes0.54
Hashing Vectorizer and SVMNo0.73
Yes0.66
Hashing Vectorizer and Logistic RegressionNo0.73
Yes0.57
TF-IDF and Linear SVMNo0.72
Yes0.54
TF-IDF and SVMNo0.70
Yes0.66
TF-IDF and Logistic RegressionNo0.73
Yes0.56
Spark NLP (Universal Sentence Encoder and Deep Learning Classifier)No0.72
BERT-base (uncased)No0.69
XLNet (base, epochs=10)No0.71
XLNet (large, epochs=10)No0.71
facebook/bart-large-mnli (fine tuned)No0.75
From the predicted stance, we notice that Democrats favored most of the topics, apart from Capitol Hill Incident and Build Back Express Tour, where they had a neutral stance. However, Republicans had a favorable outlook for 14 topics, an unbiased view for seven (The Trump & Biden Records, National Security, Leadership, Black Lives Matter, Capitol Hill Incident, Inauguration Ceremony, and Build Back Express Tour), and they were against one topic (Trump Healthcare Plan). Additionally, the predicted stance illustrates symmetry between the candidates’ tweets and the tweets from their political parties.

Conclusion & future work

Social media platforms (SMPs) have evolved into strategic spaces critical to modern political campaigns. Because of the interactivity of social media, candidates can establish direct relations with their audience without an intermediary (e.g., newspapers, news channels). Not surprisingly, campaign strategists prioritize social media as the primary channel for delivering and persuading messages. Therefore, understanding how the internal co-operation of a political front helps in contesting for elections is just as important as understanding the external factors, like the candidate’s impact and the audience’s engagement. In terms of the candidate’s impact and engagement on topics, it was found that the candidate with the highest impact did not have a higher number of tweets as compared to the candidate with a lower impact. User impact ultimately depends on the topic referred to in the tweets. Also, Joe Biden and Kamala Harris understood their audience well. They tweeted about the topics receiving higher engagements than Donald Trump and Mike Pence, who were unable to identify the topics of social interest. Additionally, the Democratic candidates displayed higher internal cooperation through their tweets and stance on different topics than the Republican candidates. This study extends political campaign research by investigating two broad aspects — engagement and stickiness of topics, the candidate synergy (i.e., the content and stance similarity) and their social reach on Twitter during the 2020 US Presidential elections. The results clearly indicate that the tweet’s topic, and the candidate’s influence, both have an impact on the amount of public engagement it receives. Internal cooperation (i.e., similarities in the content and stance for specific topics) also helps the candidates in creating a stronger hold on the thoughts of the public during the election campaign. Furthermore, this study employed an empirical approach by examining how internal cooperation between candidates and their engagement influenced the election results and it may be utilized to create a Twitter communication model for the candidates and government to assist in effective campaigning and governance. Several restrictions, as well as recommendations, must be noted for further research. Additionally, as the data for this study are focused on the tweets from Presidential and Vice-Presidential candidates for the 2020 US Elections, it was not feasible to distinguish between organic and synthetically generated (i.e., through bots, masked profiles) engagements. The impact of moving window for the content similarity between two candidates needs to be investigated further. Also, we do not intend to generalize our results for every political campaign. Moreover, our research is based on text features; we could not account for the influence of image features and knowledge graphs related to a particular tweet. Investigating the separation and impact of organic and sponsored audience involvement may benefit future studies. Additionally, factors that might influence public engagement and internal cooperation, such as emotions, timing impacts, and tweet formats, should be examined precisely. Furthermore, investigating the extent of interactions between users and political candidates through their social media managers will contribute significantly to a better understanding of the democratic ability of SMPs.
  10 in total

1.  The effects of reputational and social knowledge on cooperation.

Authors:  Edoardo Gallo; Chang Yan
Journal:  Proc Natl Acad Sci U S A       Date:  2015-03-09       Impact factor: 11.205

2.  Self-regulation versus social influence for promoting cooperation on networks.

Authors:  Dario Madeo; Chiara Mocenni
Journal:  Sci Rep       Date:  2020-03-16       Impact factor: 4.379

Review 3.  Ingroup favoritism in cooperation: a meta-analysis.

Authors:  Daniel Balliet; Junhui Wu; Carsten K W De Dreu
Journal:  Psychol Bull       Date:  2014-09-15       Impact factor: 17.737

4.  The rise and fall of cooperation through reputation and group polarization.

Authors:  Jörg Gross; Carsten K W De Dreu
Journal:  Nat Commun       Date:  2019-02-15       Impact factor: 14.919

5.  Ideological differences in engagement in public debate on Twitter.

Authors:  Felix Gaisbauer; Armin Pournaki; Sven Banisch; Eckehard Olbrich
Journal:  PLoS One       Date:  2021-03-25       Impact factor: 3.240

6.  A performance comparison of supervised machine learning models for Covid-19 tweets sentiment analysis.

Authors:  Furqan Rustam; Madiha Khalid; Waqar Aslam; Vaibhav Rupapara; Arif Mehmood; Gyu Sang Choi
Journal:  PLoS One       Date:  2021-02-25       Impact factor: 3.240

7.  #Election2020: the first public Twitter dataset on the 2020 US Presidential election.

Authors:  Emily Chen; Ashok Deb; Emilio Ferrara
Journal:  J Comput Soc Sci       Date:  2021-04-02

8.  AdCOFE: Advanced Contextual Feature Extraction in conversations for emotion classification.

Authors:  Vaibhav Bhat; Anita Yadav; Sonal Yadav; Dhivya Chandrasekaran; Vijay Mago
Journal:  PeerJ Comput Sci       Date:  2021-12-09

9.  (Re)shaping online narratives: when bots promote the message of President Trump during his first impeachment.

Authors:  Michael C Galgoczy; Atharva Phatak; Danielle Vinson; Vijay K Mago; Philippe J Giabbanelli
Journal:  PeerJ Comput Sci       Date:  2022-04-15
  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.