Literature DB >> 28363883

Methods for Coding Tobacco-Related Twitter Data: A Systematic Review.

Brianna A Lienemann1, Jennifer B Unger1, Tess Boley Cruz1, Kar-Hai Chu1.   

Abstract

BACKGROUND: As Twitter has grown in popularity to 313 million monthly active users, researchers have increasingly been using it as a data source for tobacco-related research.
OBJECTIVE: The objective of this systematic review was to assess the methodological approaches of categorically coded tobacco Twitter data and make recommendations for future studies.
METHODS: Data sources included PsycINFO, Web of Science, PubMed, ABI/INFORM, Communication Source, and Tobacco Regulatory Science. Searches were limited to peer-reviewed journals and conference proceedings in English from January 2006 to July 2016. The initial search identified 274 articles using a Twitter keyword and a tobacco keyword. One coder reviewed all abstracts and identified 27 articles that met the following inclusion criteria: (1) original research, (2) focused on tobacco or a tobacco product, (3) analyzed Twitter data, and (4) coded Twitter data categorically. One coder extracted data collection and coding methods.
RESULTS: E-cigarettes were the most common type of Twitter data analyzed, followed by specific tobacco campaigns. The most prevalent data sources were Gnip and Twitter's Streaming application programming interface (API). The primary methods of coding were hand-coding and machine learning. The studies predominantly coded for relevance, sentiment, theme, user or account, and location of user.
CONCLUSIONS: Standards for data collection and coding should be developed to be able to more easily compare and replicate tobacco-related Twitter results. Additional recommendations include the following: sample Twitter's databases multiple times, make a distinction between message attitude and emotional tone for sentiment, code images and URLs, and analyze user profiles. Being relatively novel and widely used among adolescents and black and Hispanic individuals, Twitter could provide a rich source of tobacco surveillance data among vulnerable populations. ©Brianna A Lienemann, Jennifer B Unger, Tess Boley Cruz, Kar-Hai Chu. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 31.03.2017.

Entities:  

Keywords:  Internet; review; social marketing; tobacco

Mesh:

Year:  2017        PMID: 28363883      PMCID: PMC5392207          DOI: 10.2196/jmir.7022

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   5.428


Introduction

As Twitter has grown in popularity to 313 million monthly active users [1], researchers have increasingly been using it as a data source for tobacco-related research. Twitter is a microblogging platform where users have 140 characters to share thoughts, jokes, information, images, and URLs (ie, Web addresses). Twitter posts (ie, tweets) are in real time and often public, with the potential to reach a wide audience. Users can retweet or share tweets with others, which can cause tweets to spread to large numbers of users (“go viral”). Posts can be seen when users follow each other or search for specific terms (eg, #vape). Hashtags signify a topic for users to participate in the conversation. Antitobacco campaigns may use a hashtag to start a conversation about the harms of tobacco, for example, the Truth Initiative account, @truthinitiative, promotes the use of #tobaccofreegen in the user description [2]. Similarly, the tobacco industry and independent manufacturers can use Twitter to advertise their products. Imperial Brands uses such an approach when it promotes the electronic cigarette (e-cigarette) blu through its account, @blucigs, with the hashtag #JustYouAndblu in the user description and messages to engage with Twitter users [3]. Furthermore, social media can be used to counter antitobacco campaigns. When the anti–e-cigarette campaign Still Blowing Smoke [4] was released by the State of California on television and Facebook, the pro–e-cigarette campaign Not Blowing Smoke [5], developed by pro-vaping groups, countered by taking the Twitter handles @StillBlwngSmoke [6] and @NotBlwngSmoke [7] to challenge the California campaign’s messages [8]. Tobacco-related tweets can reach a relatively young and ethnically diverse audience. Pew Research Center studies have found that, in the United States, a third of teenagers use Twitter [9], while 20% of adults have accounts with the majority being younger than 50 years [10]. Twitter is especially popular among girls aged 15 to 17 years with 49% having accounts [9]. Similarly, 45% of black, 34% of Hispanic, and 31% of white adolescents use Twitter [9]. Comparably, a larger proportion of black (28%) and Hispanic (28%) than white (20%) adult Internet users also use Twitter [10]. Therefore, tobacco conversations on Twitter, whether pro or anti, may be particularly likely to reach these populations. Research utilizing Twitter data is fairly novel without established standards across studies. Thus, it could be advantageous to establish what methods are being used and their strengths and weaknesses. Standards for reporting social media data are needed to be able to compare methods and results across studies [11]. This review focuses specifically on the methodology of tobacco-related studies that code Twitter data categorically by examining data collection methods, coding methods, and coding categories. It addresses the questions, “What methodologies are used to categorically code tobacco-related Twitter data” and “What recommendations can be made for future studies?”

Methods

Data Sources

A literature search was conducted in July 2016 using the databases PsycINFO, Web of Science, PubMed, ABI/INFORM, Communication Source, and the journal Tobacco Regulatory Science. Searches included a Twitter term and a tobacco term: (Twitter OR tweet) AND (tobacco OR nicotine OR...) (Table 1). Tobacco terms were selected based on an article on noncigarette tobacco products [12] and the US Food and Drug Administration’s (FDA) article, Recognize Tobacco in Its Many Forms [13]. Searches were limited to peer-reviewed journals in English published from January 2006 to those available in July 2016. The beginning date was selected because Twitter was launched in 2006. The initial search produced 274 nonduplicate articles (Figure 1).
Table 1

Tobacco search terms.

Search termaTobacco products covered by search term
tobaccoTobacco, smokeless tobacco, chewing tobacco, dissolvable tobacco
nicotineNicotine, electronic nicotine delivery system
cig*Cigarette, cigar, little cigar, large cigar, cigarillo, electronic cigarette, e-cigarette, e-cig
pipePipe, waterpipe
bidiBidi
kretekKretek
shishaShisha
hookahHookah, e-hookah, hookah pen
narghileNarghile
argilehArgileh
cherootCheroot
smok*Smoke, smokeless tobacco, smoking, smoker
chewChew, chewing tobacco
snuffSnuff, dry snuff, moist snuff
snusSnus
betel quidBetel quid
gutkhaGutkha
zardaZarda
toombakToombak
dissolvableDissolvable, dissolvable tobacco
ENDSENDS (electronic nicotine delivery system)
vap*Vape, vaper, vape pen, vaping, vapor

aAsterisk (*) represents stemmed words; for example, cig* would capture all words beginning with cig.

Figure 1

PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) diagram of articles included in the systematic review.

Tobacco search terms. aAsterisk (*) represents stemmed words; for example, cig* would capture all words beginning with cig. PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) diagram of articles included in the systematic review.

Study Selection

One coder independently reviewed all titles and abstracts and selected 43 articles from the initial 274 that referenced any of the tobacco products and Twitter. Full text of the articles was then reviewed. Explicit inclusion criteria were determined a priori to reduce coder bias when selecting articles. Articles were included in the review if they met the following criteria: (1) original research in a peer-reviewed journal, (2) focused on tobacco, (3) analyzed Twitter data, and (4) coded data categorically (eg, sentiment, theme). If an article was excluded for failing to meet multiple inclusion criteria, it was counted in the higher-order criteria for exclusion. A total of 16 articles were excluded, so 27 articles were included in the review (Figure 1). We excluded 1 article because it was a narrative review rather than original research, 2 articles because they did not address a tobacco product, 1 article because it analyzed Web search results and tobacco control websites rather than Twitter data, and 12 articles because they did not categorically code tobacco-related Twitter data. For example, a study may have tracked changes in the number of tweets over time but not coded the tweets categorically. The 27 included articles ranged from 2011 to 2016. Although there is risk for bias in studies, this review considered all studies that met the inclusion criteria to evaluate the strengths and weakness of all methodological approaches within this domain. The primary focus of this review is the methodology of qualitative studies rather than the outcomes of quantitative studies. Therefore, the risk of publication bias of studies with significant results and selective reporting of significant results is minimal for this type of review compared with a meta-analysis of quantitative outcomes.

Data Extraction

One author read each article to extract methodological information on data collection (data sources, date collected, tobacco topics, search keyword selection), coding methods (type of coding method, number of coders, number of tweets coded, coded retweets, number of Twitter accounts, followed URLs, coding agreement), and coded categories. This information is presented in detail in Tables 2, 3, and 4 and is summarized in the results.
Table 2

Data collection methods.

ArticleDate collectedType of tweets or accountsKeyword selectiona,bData sourceRetrieval precisionRetrieval recall
[8]March 22 to June 27, 2015Tweets about the California Department of Public Health “Still Blowing Smoke” media campaign about the harms of e-cigarettes and the pro–e-cigarette campaign “Not Blowing Smoke”#stillblowingsmoke, stillblowingsmoke, “still blowing smoke”, stillblngsmoke, “still blng smoke”, #notblowingsmoke, notblowingsmoke, “not blowing smoke”, notblngsmoke, “not blng smoke”, @CAPublicHealthGnip97.5%NRc
[25]May 1, 2013, to May 1, 2014E-cigarettesvaping, vape, vaper, vapers, vapin, vaped, evape, vaporing, e-cig*, ecig*, e-pen, epen, e-juice, ejuice, e-liquid, eliquid, cloud chasing, cloudchasing, deeming AND regulation, deeming AND FDA, deemed AND FDA, deem* AND FDAGnip59.23%NR
[26]July 1, 2008, to February 28, 2013E-cigarettes55 keywords (only examples reported): general e-cigarette terms (eg, electronic cigarette, eCig), specific brand names (eg, blu, NJoy, green smoke), and terms about e-cigarette use (eg, vaping) Excluded words related to tobacco or drugs (only examples reported): marijuana, hookahRadian691%93%
[27]December 5, 2011, to July 17, 2012 (15-day intervals)Tobaccocig*, nicotine, smok*, tobacco; hookah, shisha, waterpipe, e-juice, e-liquid, vape and vapingTwitter’s Streaming APId57.25%95%-99%
[28]January 1 to December 31, 2014E-cigarettes and smoking cessationE-cigarette keywords: vaping, vaper, vapes, vapers, e-cigarette, e-cig, mod, eGo, mod, RBA, RDA, vape, “vape pen”, “e-hookah”, “e-pipe”, “e-shisha”, “hookah pen”, “vape pipe”, #vape #vapelife #vapor #vapeporn #vapenation #vapestars #vaperazzi #vapingstyle #vaperevolution #vapeswag #vapesirens #vaperscommunity #vapepics #vapesociety #socialvapers #vapefamily #vapefreedom #vapelove #vapers #vapstagram #vapelyfe #vapeshop #vapeon #vapestrong #girlswhovape #alldayvape #adv #vapersoul #VGOD #ecig #ecigarette Smoking cessation keywords: quit, stop, “quit smoking”, “stop smoking”, “quit cigarettes”, “smoke less”, “smoking less”, help, NRT, patch, lozenge, spray, gum, nicorette, nicotine, #quit #quitsmoking #quitsmokingcigarettes #Cessationnation Exclusion keywords: marijuana, weed, pot, dank, trees, green, cheeba, THC, cannabis, sativa, indica, bud, marihuana, MJ, “mary jane”Sysomos HeartbeatNRNR
[29]February 1 to April 30, 2014Blu e-cigarettes’ tweets and retweets@blucigsTwitter REST APINRNR
[30]April 12 to May 10, 2014Hookah or shishahookah, #hookah, shisha, #shisha, hooka, #hooka, sheesha, #sheeshaSimply Measured99.56%NR
[31]November 1, 2011, to August 31, 2013Hookah, cigarettes, and cigarscigar, cigars, cigarette, cigarettes, hookah, waterpipe, water pipe, shisha, sheeshaTwitter’s Streaming APINRNR
[32]October 4 to November 3, 2010TobaccoSmoking, tobacco, cigarette, cigar, hookah, hookaTwitter’s Search APINRNR
[33]May 1, 2012, to June 30, 2012E-cigarettesKeywords: e-cigarette, ecigarette, e-cig, ecig Additional keywords AND “cig” or “cigarette”: electronic, blu, njoyGnip>99% of a random sample of 500 tweetsNR
[34]December 6, 2012, to June 20, 2013Tobacco or cessation price promotionTobacco-related: cig(s), cigarette(s), smoking, tobacco, blu cigarette, njoy cigarette, ecig, e-cig, @blucig, e-cigarette, ecigarette, from:blucigs*, ecigs, e-cigs, ecigarettes, e-cigarettes, “green smoke”, “south beach smoke”, cartomizer, (atomizer OR atomizers) –perfume*, ehookah OR e-hookah, ejuice OR ejuices OR e-juice OR e-juices, eliquid OR eliquids OR e-liquid OR e-liquids, e-smoke OR e-smokes, (esmoke OR esmokes), eversmoke, “joye 510”, joye510, lavatube OR lavatubes, logicecig OR logicecigs, smartsmoker, smokestik OR smokestiks, “v2 cig” OR “v2 cigs” OR v2cig OR v2cigs, vaper OR vapers OR vaping, zerocig OR zerocigs, cartomizers, Vuse, MarkTen Price-related: Coupon(s), Promo(s), Promotions(s), Promotional, Discount(s)(ed), Save, Code(s)Gnip56.94%NR
[35]July 2014Slogans for the Dutch health campaign “Smoking is so outdated” (Roken kan echt niet meer)#rokenkanechtnietmeer [#smokingissooutdated]Twitter’s Search APINRNR
[36]December 2013Little cigarsSwisher Sweets, Black & MildsTwitonomy67.50%NR
[37]September 2012 and January to May 2013Genetic information on smokinggenetic, smokingNR49.1%NR
[38]August 2010Smoking cessation accountsSearched for smoking cessation accounts using the following terms: “quit or stop smoking” or “smoking cessation”NRNRNR
[39]January 8-15, 2014Tweets about Chicago Department of Public Health’s e-cigarette Twitter campaign@ChiPublicHealthtwitteR package for R and NodeXLNRNR
[40]January 2010 to January 2015E-cigarettesvape, ecig, ecigarette, vaping, ejuice, vapers, drip AND tip, dripping, eliquid AND flavor, e AND juice, e AND liquid, smoke AND free, off AND cigarettes, ex AND smoker, no AND analogs, I AND quitNRNRNR
[41]January 2012 to December 2014E-cigarettese(-)cig, e(-)cigarette, electronic cigarette, etcTwitter’s Streaming API81% to 90.8% for 4 groups of 500 randomly sampled automated tweetsNR
[42]September to December 2013 and March 2015E-cigarettesElectronic-cigarette, e-cig, e-cigarette, e-juice, e-liquid, vape-juice, vape-liquidTwitter’s Streaming API and Twitter’s Firehose97.21%86.63%
[43]April 21 to October 20, 2014Blu and V2 e-cigarettes’ tweets and retweets@blucigs, @v2cigsNR100%NR
[44]July 7 to 21, 2014Tweets about the Centers for Disease Control and Prevention’s (CDC) Tips From Former Smokers campaign#cdctips, CDC AND smokingSocial Network Importer for NodeXL81.70%NR
[45]May 1, 2013, to May 1, 2014E-cigarettesvaping, vape, vaper, vapers, vapin, vaped, evape, vaporing, e-cig*, ecig*, e-pen, epen, e-juice, ejuice, e-liquid, eliquid, cloud chasing, cloudchasing, deeming AND regulation, deeming AND FDA, deemed AND FDA, deem* AND FDAGnip59.23%NR
[46]March to June 2013Tobacco control program tweets during the months that the national CDC Tips smoking cessation campaign airedGoogle search for tobacco control programs using the terms “ tobacco program ” and “ quitline .” If the site included a link to a Twitter account, that account was included.Radian6NRNR
[47]March 15 to June 9, 2012Tweets about the CDC’s Tips campaignTobacco behavior: cig(s), cigarette(s), nicotine, smoke(s), smoker, smoking, tobacco Tobacco policy: @cdcgov, @cdctobaccofree, @drfriedencdc, @fdatobacco, @smokefreegov, antitobacco, antismoking, CDC, quitline, quitnow, secondhand+smoke, smokefree, smokefree.gov, tobaccofree Ad specific: #cdctips, amputation, amputee, Buerger’s+Disease, heart+attack, hole+neck, hole+throat, lung+cancer, stoma, stroke, throat+cancer Engagement: ad, commercial, campaign, PSAGnip78.87%94%
[48]February 5-12, 2014CVS Health-related tweets surrounding the announcement of ending tobacco sales#cvs, #cvsquitsTwitter’s Streaming API72.38%NR
[49]50 most recent tweets from July 18, 2012Smoking cessation accountsSearched for smoking cessation accounts using the terms: “quit smoking” and “smoking cessation”NRNRNR
[50]February 23 to April 9, 2015Exposure to secondhand e-cigarette aerosol“secondhand vape” OR “secondhand vaping” OR “second-hand vape” OR “second-hand vaping” OR “vape smoke” OR “ecig smoke” OR “e-cig smoke” OR “e-cigarette smoke” OR “vape shs” OR “ecig shs” OR “vape secondhand smoke” OR “vape second-hand smoke” OR “esmoke” OR “e-smoke”NCaptureNRNR

aAsterisk (*) represents stemmed words; for example, cig* would capture all words beginning with cig.

bWords in italics were not keywords used for searches.

cNR: not reported.

dAPI: application programming interface.

Table 3

Coding methods.

ArticleCoding methodNo. of codersNo. of tweets codedCoded retweetsNo. of Twitter accountsFollowed URLsCoding agreement
[8]Hand-coded by researchers1: all tweets; 2: subsample 300 tweets2248: relevance; 2192: contentYesNRaNo91%: sentiment; 72%: theme
[25]Hand-coded by researchers6: for a subset of 250 tweets; NR for total17,098: relevance; 10,128: contentYes, if additional contextNRYesκ=.64 to .70
[26]Machine learning with initial hand-coding; Python Scikit-LearnNR1,669,123YesNRYesNR
[27]Machine learning and hand-coding; naïve Bayes, k-nearest neighbors, and support vector machines2: pilot of 1000; 2: random subset of 150; 2: all 73627362: relevance; 4215: contentRetweeted posts were onlyincluded onceNRNRκ>.70 for the random subset of 150
[28]Hand-coded by researchers1: all tweets; 2: for 10% subsample300: complete sample; 300: industry-free sample; 481 of 600: content (duplicates between samples removed)Yes148: complete sample; 215: industry-free sampleYesκ=.74
[29]Hand-coded by researchers2NRYesApproximately 3400NRNR
[30]Crowdsourcing with initial hand-coding35000: relevance; 4978: contentNR3804NRκ=.66 to .85 among a subset coded by researchers
[31]Topic modeling with machine learning; MALLET, a command-line implementation of latent Dirichlet allocation (LDA)NR319,315: total; 95,738: hookah; 22,513: cigar; 201,064: cigaretteNRNRNRNR
[32]Topic modeling (LDA) with machine learningNR4962NRNRNRNR
[33]Machine learning and hand-coding; DiscoverText2: for a subset of 500 for relevance, 4500 for commercial versus organic, 7500 for cessation73,672Yes23,700Yes, hand-coded tweets with URLsκ=.87 to .93
[34]Hand-coded by researchers1: all; 2: for subsets of 100 tweets5000: relevance; 2847: contentNRNRYesκ=.64 to 1.00
[35]Hand-coded by researchers1: all tweets; 3: subsample133NoNRNRalpha = .61 to 1.00
[36]Hand-coded by researchers33935: relevance, foreign language, retweets; 2656 sampled for 288 original tweets for codingNo346Yesκ=.64 to .91
[37]Hand-coded by researchers; wordcloud R packageNR171: relevance; 84: contentNR84NRNR
[38]Hand-coded by researchers1: all tweets; 2: for 20% of tweets143,287: identified; 4753: coded for clinical practice guidelines for treating tobacco dependenceNR153Yes>90%
[39]Hand-coded by researchers2684Yes306YesNR
[40]Machine learning and hand-coding; naïve Bayes, LIBLINEAR, Bayesian logistic regression, random forests; keyword comparisons1: all tweets; 2: subsample of 200013,146NR2147No, removed URLsκ=.87 for subsample
[41]Machine learning and hand-coding; human detection algorithm; Hedonometrics; key phrasal pattern matching2: for all tweets from 500 automated accounts and 500 organic accounts as classified by the algorithm; 2: for 4 groups of 500 randomly sampled tweets to gauge accuracy of subcategorical tweet topics850,000Yes131,622: automated accounts; 134717: organic accounts: 188,182: not classified accounts (ie, accounts with <25 tweets)No, but the algorithm used the count of URLs to distinguish automated accounts from organic accounts; also used keywords in the URLs for the algorithm to determine subcategories of automated accounts94.6% true- positive rate, 12.9% false- positive rate for the machines on the tweets from the 1000 accounts also coded with human-coding
[42]Machine learning with initial hand-coding; Python Scikit-Learn; topic modeling with MALLET2: for a subset of 1000 profiles224,000 in 2013 sample; 349,401 in 2015 sampleYes34,000 in 2013 sample; 100,000 in 2015 sampleNo; metadata on the presence of URL linksκ=.88
[43]Hand-coded by researchers and MySQL pattern matcherNR1180Yes2: Blu and V2; 537: users retweeting Blu and V2NRNR
[44]Hand-coded by researchers1: all tweets; 2: for 20% of tweets (n=358)2191: relevance; 1790: contentYesNR (>21)NRκ=.95 for 20% subsample
[45]Machine learning with initial hand-coding; naïve Bayes classifier, k-nearest neighbors, support vector machines6: for a subset of 250 tweets; NR for total17,098: relevance; 10,128: contentYes, if additional contextNRNRκ=.64 to .70
[46]Hand-coded by researchers31776No16YesFor 5% of data, 95.7%; κ=.72
[47]Machine learning with initial hand-coding; naïve Bayes classifier2: subset of 450 tweets for relevance; 2: subset of 350 tweets for content245,319: relevance; 193,491: contentNR166,857NR; metadata on the presence of URL linksκ=.93
[48]Hand-coded by researchers1: all tweets; 2: for 1% of tweets8645: relevance; 6257: contentYesNRYes90% for a 1% sample of tweets
[49]Hand-coded by researchers2900, with 50 tweets per accountYes18NR84%
[50]Hand-coded by researchers21519No1321Yesκ=.84

aNR: not reported.

Table 4

Coded categories.

Category typeCategoryNumber of articles and percent of totala n (%)Articles
RelevanceRelevant versus nonrelevant16 (59)[8,25,27,30,33,34,36,37,41-45,47-49]
Sentiment9 (33)[8,25,27,28,35,41,45,48,50]
Positive or negative (ie, supportive or against)6 (22)[8,25,27,45,48,50]
Positive or negative (ie, emotional tone)2 (7)[28,41]
Positive or negative valence1 (4)[35]
Neutral or unknown6 (22)[8,27,28,45,48,50]
Message attitudePro or con1 (4)[28]
Type of utteranceComparison versus attribution versus metonymy1 (4)[35]
Topics, themes, or genres21 (78)[8,25-28,30,33,34,36-43,45-47,49,50]
Joke or humorous3 (11)[27,28,36]
Song or music2 (7)[30,36]
Profanity1 (4)[36]
Social relationships2 (7)[27,50]
Sex or romance1 (4)[30]
Image or stereotype1 (4)[27]
Risky behaviors or other substances6 (22)[25,27,28,30,36,45]
Illicit substance use in e-cigarettes2 (7)[25,45]
Preference for another substance1 (4)[30]
Affiliation and preference1 (4)[36]
Flavors7 (26)[25,36,39,41-43,45]
Pleasure1 (4)[27]
Tastes good1 (4)[28]
Craving, desire, and need5 (19)[25,27,28,36,45]
Addiction1 (4)[37]
Type of tobacco product4 (15)[27,30,34,38]
Type of tobacco product brand1 (4)[26]
E-cigarettes’ smoke-free aspect1 (4)[42]
Health, safety, harms9 (33)[8,25,27,30,33,37,39,45,50]
Downplayed or refuted harms, harm reduction2 (7)[27,42]
E-cigarettes for smoking cessation7 (26)[25,28,33,40-42,45]
Cessation5 (19)[27,30,37,41,46]
Cessation product2 (7)[34,38]
Socioemotional support tweets regarding quitting smoking1 (4)[49]
Encouraging or engaging tweets regarding quitting smoking1 (4)[49]
Clinical practice guidelines for treating tobacco dependence1 (4)[38]
Demonstration1 (4)[36]
Use8 (30)[25,27,28,30,36,37,40,45]
Use: general2 (7)[36,40]
First-person use or intent5 (19)[25,27,28,30,45]
Second- or third-person experience4 (15)[25,27,28,45]
Starting use or smoking initiation3 (11)[27,28,37]
Recent use1 (4)[30]
Underage use3 (11)[25,27,45]
Parental use2 (7)[25,45]
Does not use or does not want to use1 (4)[30]
Secondhand smoke1 (4)[46]
Rejection and prevention1 (4)[36]
Disgust, unattractive, or uncool2 (7)[27,30]
Policy, government, regulation, activism, politics7 (26)[8,25,27,28,39,45,46]
Normalization versus discouragement1 (4)[30]
Getting others started or advocating use1 (4)[28]
Attempt to engage other Twitter users1 (4)[28]
Fear appeals1 (4)[47]
Lies or propaganda2 (7)[8,39]
Advertisement, promotion, marketing, industry, commercial12 (44)[8,25-28,30,33,36,38,41,45,50]
Offering advice1 (4)[28]
Personal opinion or communication6 (22)[25,27,28,33,38,45]
News or update4 (15)[25,27,28,45]
Information5 (19)[25,27,28,45,49]
Science or scientific publication2 (7)[37,39]
Cultural reference1 (4)[27]
Issue salience1 (4)[39]
Commodity1 (4)[27]
Connoisseurship1 (4)[27]
Cheaper than smoking1 (4)[28]
Money1 (4)[8]
Price promotion, discount, coupon4 (15)[26,33,34,41]
Backgrounded1 (4)[27]
Other or undetermined2 (7)[36,50]
Domains smoking was compared with for campaign slogansPersonal features; hobby or hype; person or group; social norm; big event; technology and innovation; sex or relation; eating, drinking, and stimulants; school; transport; and campaign1 (4)[35]
Links (URLs)Most common links1 (4)[26]
Location of use1 (4)[26]
Class1 (4)[26]
House, room, bed1 (4)[26]
School1 (4)[26]
Public1 (4)[26]
Bathroom1 (4)[26]
Work1 (4)[26]
In front of someone1 (4)[26]
Car1 (4)[26]
Restaurant1 (4)[26]
Movie theater1 (4)[26]
Airplanes or airport1 (4)[26]
Store1 (4)[26]
Bars or clubs1 (4)[26]
Dormitory1 (4)[26]
Library1 (4)[26]
Mall1 (4)[26]
Bowling alley1 (4)[26]
Café or coffee shop1 (4)[26]
Hospital1 (4)[26]
Locker room1 (4)[26]
Topic modeling2 (7)[31,32]
Hookah topic 1: social locations, leisure time, and positive affect1 (4)[31]
Hookah topic 2: fun, leisure time, and sociability1 (4)[31]
Cigarette topic 1: death and unpleasant smell1 (4)[31]
Cigar topic 1: positive affect and enjoyment1 (4)[31]
Cigar topic 2: luxury alcohol products1 (4)[31]
Tobacco topic 1: tobacco use and substance use1 (4)[32]
Tobacco topic 2: addiction recovery1 (4)[32]
Tobacco topic 3: addiction recovery and tobacco promotion by clubs or bars1 (4)[32]
Tobacco topic 4: tobacco promotion by bars or clubs and marijuana use1 (4)[32]
Tobacco topic 5: antismoking and addiction recovery1 (4)[32]
User or account10 (37)[8,25,28,29,34,37,41,44,45,49]
Government3 (11)[25,44,45]
Foundations or nonprofit organizations4 (15)[25,44,45,49]
Public health and health care1 (4)[28]
Researcher or research center2 (7)[29,37]
News5 (19)[25,28,37,44,45]
Reputable news source2 (7)[25,45]
Press, media, or news3 (11)[28,37,44]
Medical news source1 (4)[37]
Personal accounts7 (26)[8,25,28,29,37,44,45]
Personal accounts, everyday people, individuals6 (22)[8,25,28,37,44,45]
Personal accounts with industry ties1 (4)[28]
Person: supporter1 (4)[29]
Person: basic profile (no mention of e-cigarettes)1 (4)[29]
Celebrity, public figures3 (11)[25,28,45]
Organic (human)1 (4)[41]
E-cigarette community movement2 (7)[25,45]
Industry5 (19)[25,28,29,34,45]
Industry: retailer or manufacturer2 (7)[28,29]
Retailer or vendor3 (11)[25,34,45]
Tobacco company2 (7)[25,45]
Industry: other (eg, vaping magazine, Web marketer)1 (4)[29]
For-profit organization1 (4)[44]
Entity: general (eg, company, store, advocacy group)1 (4)[8]
Nonperson (eg, musical band)1 (4)[29]
Bots, automatic, fake4 (15)[25,28,38,41,45]
Unclassified or other5 (19)[8,29,37,44,49]
Profile photo1 (4)[36]
Single person versus multiple people1 (4)[36]
Gender (male, female, mixed group)1 (4)[36]
Age (babies or children, high school or college, adult)1 (4)[36]
Race (African American, white, Hispanic, Asian, undetermined)1 (4)[36]
Location of user4 (15)[8,37-39]
City, state, and country1 (4)[39]
State1 (4)[8]
Country1 (4)[37]
Continent1 (4)[38]

aPercentages are rounded to the nearest whole percent.

Results

Data Collection Methods

Data Sources

In total, 22 of the 27 articles discussed the sources they used for their sample of Twitter messages, while 5 articles did not. Twitter provides 3 primary sources of data: Twitter’s Search application programming interface (API), Twitter’s Streaming API, and Twitter’s Firehose. An additional method is Twitter’s REST API, which allows tracking specific users by their username. One study collected data using Twitter’s REST API. Twitter’s Search API is free to use and provides a maximum of 3200 past tweets (ie, published in the past 7 days, so it is not in real time) with a limit of 180 searches every 15 minutes [14,15]. Twitter’s Search API was used in 2 studies (Table 2). There are programs developed to interact between Twitter’s Search API and specific analyses programs. For example, twitteR package for R [16], NCapture for NVivo [17], and Social Network Importer for NodeXL, the free add-on for Excel [18], provide access to Twitter’s Search API. A small subset of the studies used these programs: twitteR package for R (n=1), NCapture (n=1), and Social Network Importer for NodeXL (n=1). More useful to researchers is Twitter’s Streaming API, which provides all tweets related to the search terms up to a limit of 1% of the Twitter database for that time period. If the selected search terms are infrequently used across Twitter (eg, the name of a local tobacco campaign), all tweets related to the search terms will be available. However, if the selected search terms are commonly used, tweets related to those terms will be available up to a limit of 1% of the current Twitter database. Twitter’s Streaming API is free, publicly available data [19]. A total of 5 studies used Twitter’s Streaming API (Table 2). Twitonomy is an analytics tool that accesses Twitter’s Streaming API and offers both free and premium packages for a small fee [20]. One article used Twitonomy. Providing the greatest access to data, Twitter’s Firehose has real-time access to 100% of Twitter content. Twitter’s Firehose formerly was handled by multiple data providers (eg, Gnip, DataSift, and Topsy). After the acquisition of Gnip in 2014, Twitter transitioned to only allowing access to Twitter’s Firehose through Gnip for a fee from August 2015 onward [21]. A total of 6 articles used Gnip, making it the most common method to collect data (Table 2). There are social listening programs that buy Twitter data from data providers such as Gnip. Radian6 [22], Simply Measured [23], and Sysomos Heartbeat [24] collect data from across social networking sites, blogs, forums, and news sites for a fee. A few articles used these social listening programs to collect data: Radian6 (n=2), Simply Measured (n=1), and Sysomos Heartbeat (n=1). Data collection methods. aAsterisk (*) represents stemmed words; for example, cig* would capture all words beginning with cig. bWords in italics were not keywords used for searches. cNR: not reported. dAPI: application programming interface.

Date Collected

Twitter data were collected across the studies from 2008 to 2015. The time span of Twitter data collected within an individual study ranged from 1 day to 5 years with a median of 14 weeks (Table 2).

Tobacco Topics

Among the 27 studies, 41% (n=11) analyzed messages related to e-cigarettes, 19% (n=5) related to other tobacco products, and 22% (n=6) about specific tobacco campaigns (Table 2). The remaining topics included smoking cessation accounts (n=2), tobacco or cessation price promotion (n=1), genetic information on smoking (n=1), and ending tobacco sales at CVS Health (n=1).

Search Keyword Selection

Kim and colleagues [11] proposed a framework of three steps to develop and validate search filters. This framework was selected because it provided a standard in which to compare studies. Most of the articles partially fulfilled these steps within the framework. The first step is to develop a search filter. All the articles generated a list of keywords presumably based on expert knowledge of the topic or a systematic search of language related to the topic of interest. However, only 4 articles discussed the process of discarding keywords that returned a high proportion of irrelevant results and adding new keywords as new terms appeared in the returned literature [25-28]. It should be noted that for some studies this process of developing a search filter may be irrelevant because they are coding all tweets from a specific account (eg, coding tweets from the blu e-cigarette account, @blucigs) [29]. Search keyword selection was tailored to the studies’ specific topics (Table 2). To search for e-cigarette Twitter data, variations on e-cigarette (eg, e-cig), vape (eg, vaping), e-liquid (eg, eliquid), and e-juice (eg, ejuice) were common. Some e-cigarette studies also included major e-cigarette companies or brands as key search terms (eg, Njoy). Studies that analyzed Twitter data on specific tobacco campaigns used a variety of tactics such as searching for variations on the campaign name (eg, still blowing smoke), the source of the campaign (eg, CDC), specific features of the campaign ads (eg, lung+cancer), tobacco products and behavior (eg, smoke), and general campaign terms (eg, PSA). Some studies also used a combination of searching for terms with and without hashtags (eg, hookah, #hookah). In 2 studies, marijuana terms (eg, weed) were used as exclusion keywords [26,28]. The second step of the framework is to apply the search filter and split data into retrieved and unretrieved sets. The third step is to assess the search filter on its ability to distinguish between relevant and irrelevant messages [11]. Precision refers to how much retrieved data are relevant, whereas recall refers to how much relevant data are retrieved. Recall is similar to measures of sensitivity. Precision is much less difficult to accurately estimate than recall because recall requires conclusions to be drawn about the tweets that were not retrieved. A precision score of 100% means that all retrieved data were relevant, while a recall score of 100% means that all relevant data were retrieved. However, precision and recall are inversely related. As a search filter expands to collect more data, the proportion of relevant data decreases. Publications can present a precision-recall curve to show the extent of this trade-off for their search filter. A good search filter will maintain a relatively high level of precision as recall increases. A total of 15 articles reported precision, which ranged from 49.1% to 100% (Table 2). A total of 4 articles reported recall, which ranged from 86.6% to 99% (Table 2).

Coding Methods

Coding methods for the studies included hand-coding, machine learning, or a combination of the two. Hand-coding involves one or more human coders categorizing data. When 2 or more coders independently code data, a coding agreement score (eg, kappa) between the coders can be calculated. In contrast, machine learning uses an algorithm for a computer to learn how to code data. However, human-coding is used for an initial subset of data to help refine the algorithm to improve its accuracy. Coding categories may be determined a priori based on prior research or they may be developed inductively through the process of coding. Studies used hand-coding only (n=17), machine learning + hand-coding (n=8), and machine learning only (n=2; Table 3). Of the studies using hand-coding, data were coded by the reviewed studies’ researchers in 16 studies, while 1 study used crowdsourcing (ie, many Web workers) [30]. All the studies that used machine learning also used initial hand-coding for a subset of the data, except for 2 studies that used topic modeling [31,32]. Topic modeling produces thematically related word clusters from the text [31]. The data collected in the hand-coded studies ranged from a collection period of 1 day to 1 year, while the machine learning studies ranged from 1 month to 5 years of Twitter data. Across the articles that used hand-coding, the number of coders per tweet ranged from 1 to 6. Coding agreement was reported by 20 articles: coding agreement percentage (n=5; 72% to 95.7%), kappa scores (n=13; kappa=.64 to 1.00), Cronbach alpha (n=1; alpha=.61 to 1.00), and both coding agreement percentage and kappa scores (n=1). Kappa values can be interpreted as poor (<.20), fair (.21-.40), moderate (.41-.60), good (.61-.80), and very good (.81-1.00) [51]. Cronbach alpha can be interpreted as unacceptable (<.50), poor (.50-.59), questionable (.60-.69), acceptable (.70-.79), good, (.80-.89), and excellent (>.90) [52]. However, it is important to note that these thresholds are not derived statistically but instead rely on intuitive judgments. The number of tweets coded per study ranged from 171 to 17,098 for hand-coding and from 7362 to 1,669,123 for machine learning. A total of 14 articles included retweets in their total number of tweets. A total of 15 articles reported the number of unique Twitter accounts, which ranged from 2 to 3804 for hand-coding studies and from 23,700 to 166,857 for machine learning studies (see Table 3). URLs in tweets can provide information that changes the context or meaning of a tweet. Following URLs to their respective webpages can be time-consuming, but it can increase coding accuracy. Machine learning algorithms can analyze the text within URLs but may require human coders to follow them to their respective webpages. A total of 15 articles reported whether they followed URLs (followed: n=10, did not follow: n=5; Table 3). One article provided the most common URLs [26]. The studies tended to show that advertising or commercial tweets were significantly more likely to contain URLs than other types of tweets [25,33,34]. Coding methods. aNR: not reported.

Coded Categories

All the studies developed categories for content. These content areas included one or more of the following: sentiment, theme, location of use, user description, profile photo, or location of user (Table 4).

Sentiment

A total of 9 articles coded for sentiment (Table 4). One article made a distinction between coding for sentiment (ie, emotional tone or affective content: positive, negative, or neutral) and message attitude (ie, pro, con, neutral or do not know) [28]. Two articles coded for sentiment in terms of emotional tone. In 6 articles, sentiment was described in terms of being supportive or against tobacco, tobacco users, or decisions regarding tobacco, which suggests an assessment of message attitude. Furthermore, 1 article assessed valence, but it was not clear whether positive or negative valence suggested an attitude or emotional tone [35].

Topic or Theme

A total of 21 studies coded for topic or theme (Table 4). The most common themes included the following: advertisement, marketing, industry or commercial (n=12 articles); health, safety, harms (n=9); use (n=8); policy, government, regulation, activism (n=7); e-cigarettes for smoking cessation (n=7); flavors (n=7); personal opinion or communication (n=6); risky behaviors or other substances (n=6); cessation (n=5); information (n=5), and craving or need (n=5). One study coded for location of use with 20 categories (eg, school, work) [26]. Finally, 2 studies used topic modeling to explore tobacco content [31,32].

User or Account

A total of 10 studies coded for user description from data found in the user profile, including type of account, age, location, and other characteristics (Table 4). The most common types of user categories coded were personal accounts (n=7 articles), industry accounts (n=5), news (n=5), unclassified (n=5), and bots, automatic, or fake (n=4; ie, automated computer program). None of the articles reported the percentage of tweets that had accounts without user profile information. However, across the articles with a category for tweets with an unclassifiable user description, 0.2%-38% of tweets were unclassified. Across the studies that coded for bots, the percentages of tweets varied drastically from 6.9% to 80.7%. One study coded Twitter profile photos with 4 categories: gender, age, race, and single person versus multiple people [36]. Visual cues (eg, skin color, background themes, facial features) in the profile pictures were used for coding. Coding for age based on available visual cues in the profile photos tended to be difficult, so coding for age was simplified to 3 broad groups: young, middle age, and older adult. There were few middle-aged and older adults represented in the sample, so a dichotomous variable of young or not young was created [36]. A total of 4 studies coded for the location of the user with one each coding for state (California vs other) [8], country [37], continent [38], and city, state, and country (United States vs other) [39]. Location was identified for 51% to 63% of Twitter profiles. Most accounts that listed a location were from the United States and North America. Coded categories. aPercentages are rounded to the nearest whole percent.

Discussion

Overview

Studies analyzing tobacco-related Twitter data have grown in number in recent years. Although we searched for articles published from 2006 to 2016, articles meeting inclusion criteria for this review were published from 2011 to 2016, with 85% (23/27) of the publications occurring in 2014-2016. Widely used among adolescents and black and Hispanic individuals, Twitter could provide a rich source of tobacco-related data among these groups. One of the benefits of Twitter research is the ability to focus on emerging issues and products that are not yet addressed in surveillance or epidemiological research. Collecting tobacco data in real time via Twitter could be a useful tool for tobacco surveillance, which could help inform tobacco control policies and social media campaigns. Twitter data can change rapidly as they are being posted daily [37]. If a study only collects data from one point in time, it may not be reflective of data at any other point in time. Twitter studies should consider collecting data at multiple time points or over longer periods of time to decrease the likelihood that results are idiosyncratic to that point in time. However, there may be some research questions that only require sampling one time or sampling directly before and after an event to gauge short-term responses. The primary sources of data were Gnip and Twitter’s Streaming API, which offer different strengths and weaknesses. If a study wants access to all tobacco-related tweets, then Gnip may be more effective. However, if a study is interested in tweets about a specific tobacco campaign or has a constrained budget, then Twitter’s Streaming API may be a better data source. Alternatively, if a study is focusing on multiple social media sites (eg, Twitter, Facebook), then using Radian6, Simply Measured, or Sysomos Heartbeat may be appealing because of their cross-platform analysis. The results of this review suggest that there are some gaps in the types of tobacco products studied by Twitter analyses. E-cigarettes were the product that was addressed by most studies, while none of the studies focused on smokeless tobacco, snus, bidis, or kreteks. The focus on e-cigarettes over other products could be due in part to their recent rise in popularity and recent debates about policies. From 2011 to 2014, e-cigarette use among high school students in the United States significantly increased from 1.5% to 13.4% [53]. Similarly, there was a significant increase in hookah use from 4.1% to 9.4%. However, there were significant decreases for cigarette (15.8% to 9.2%), snus (2.9% to 1.9%), cigar (11.6% to 8.2%), pipe (4.0% to 1.5%), and bidi (2.0% to 0.9%) use [53]. It could be useful to conduct studies comparing tweets about cigarettes, e-cigarettes, and hookah among adolescents to help understand the changing rates of use and Twitter postings that discuss issues related to use. The rates of e-cigarette, hookah, and cigar use among high school students may be affected by the FDA regulations of these products that went into effect on August 8, 2016. One of the provisions of these regulations is that it will be illegal to sell e-cigarettes, cigars, and hookah tobacco to persons younger than 18 years [54]. Researchers may want to analyze e-cigarette, cigar, and hookah tweets before and after August 8, 2016, to gain real-time insight into adolescents’ reactions to the new regulations. Future Twitter studies could benefit from a standard of reporting data collection methods. Only 4 of the articles in this review reported such rigorous methods of selecting their search keywords as suggested by Kim and colleagues’ [11] search filter framework, while 15 articles reported retrieval precision and 4 reported retrieval recall. The terminology around newer products such as e-cigarettes is growing, so it may be difficult to capture all relevant Twitter conversations with one’s keywords [25]. Future studies will need to continue to refine and expand search keywords.

Data Coding

The methods of coding were hand-coding, machine learning, or a combination of the two. Machine learning can code larger quantities of data at a quicker rate than hand-coding, but human coders may have greater discretion at coding for the complexities and subtlety of language such as humor, irony, or sarcasm. For example, algorithms developed to detect irony only retrieved 54%-57% of tweets coded as irony by multiple independent human coders [55]. Hand-coding can be subject to bias, but creating coding schemes based on prior literature and working to achieve acceptable levels of interrater reliability can help attenuate individual bias. Hand-coding allows researchers to follow URLs, which can change the meaning of the tweet. Viewing the webpage may provide additional information that may not be discernable from the URL. Studies that require determining subtle differences in context may be better suited to hand-coding a small sample of Twitter data, while studies that rely less on context could code large samples with machine learning. The sentiment of tweets could help evaluate whether the responses to pro- and antitobacco efforts are positive or negative as a way of understanding social norms about these products. Clarity and comparability across studies could be improved if a distinction is made between attitude and emotion when coding for sentiment. For example, a tweet could be pro-vaping but have a negative emotional tone or it could be anti-vaping but have a positive tone. Only 1 article clarified the meaning by making an explicit distinction between coding for emotional sentiment and message attitude [28]. This is reflective of coding for stance (in favor of, against, or neutral) versus sentiment (emotional tone) [56]. The three most common themes used for coding were advertisements or marketing; health, safety, harms; and use. Surveillance of these themes could be beneficial to understand whether tobacco advertisements are being circulated on Twitter with the potential of reaching underage individuals, whether the content is making unproven claims about the health and safety of their products, or promoting the use of their products to vulnerable populations such as youth or ethnic minorities. The most common user account descriptions that the studies coded for were personal accounts, industry, news, unclassified, and bots. Determining the user description and demographic information for the accounts that tweet about tobacco could help determine whether tobacco companies, pro-vaping advocacy groups, or antitobacco efforts are circulating with a potential to reach certain groups (eg, adolescents). Researchers could also monitor how news organizations are presenting tobacco-related information to the public. It may be difficult to determine the demographics of the person tweeting or of the audience exposed to the tweets, which could be especially problematic when studies want to focus on vulnerable populations (eg, adolescents). A Twitter account could be run by an individual, multiple people (eg, vape shop employees), or bots. To reduce bias, bots should be identified and the tweets from these accounts removed from analysis or identified as tweets originating from automated accounts [57]. Only 4 studies coded for bots with a range from 6.9% to 80.7% of tweets classified as bots. Even if an account is run by an individual, a Twitter profile provides little information. It may or may not include a photograph, profile description, location, website, and birthday. It does not include gender or ethnicity and race, so this information needs to be estimated. Although none of the studies reported the percentage of tweets with accounts missing profile information, 5 studies coded for an unclassified category with 0.2%-38% of tweets being unable to be classified based on the account profile. If profile information is included, it could be used in combination with natural language processing to infer information about the individual from his or her tweets [29]. For example, algorithms have estimated gender with a 75.5% accuracy based on tweets and a 92.0% accuracy based on tweets, screen name, full name, and profile description [58]. A study that combined analysis of text and image processing predicted gender with an accuracy of 85.1% [59]. Additionally, algorithms based on tweets were capable of predicting the exact age of the user within a margin of 4 years, while accuracy for age categories were 93.0% for <20 years, 67.4% for 20-40 years, and 81.6 for >40 years [60].

Recommendations

It is recommended that tobacco Twitter studies adopt methodological standards of reporting and data quality assessment. Important information to consider reporting include data sources (eg, Gnip, Twitter’s Streaming API), the date range of tweets collected, the number of tweets coded, whether retweets were coded, whether coders followed URLs, whether images were coded, the categories coded, the decision criteria for each category, the number of unique Twitter accounts, and the types of Twitter accounts. Studies that use hand-coding should also consider reporting the number of independent coders, the number of tweets coded across coders, and their coding agreement, while it is important for machine learning studies to detail the development and refinement of their algorithms. Providing this information is likely to increase comparability across studies and the ability to replicate results. Depending on research goals, studies may want to sample Twitter’s databases multiple times, especially if they collect from Twitter’s Search API or Streaming API, which put limitations on the amount of data that can be collected per sample. Sampling at multiple times is also important for studies that have access to Twitter’s Firehose through Gnip, considering that tweets are in real time, which could lead to fluctuations based on real-world events (eg, the FDA’s new regulations for e-cigarettes). Following messages over time could also help establish trends in the content of posts. However, some studies may be interested in short-term reactions to an event, which requires different sampling methods than following trends. For example, if a study is interested in short-term reactions to the FDA’s new regulations for e-cigarettes, then a single sample before and after the implementation of the new regulations could be sufficient. When coding for sentiment, researchers could improve clarity by making a distinction between whether they are coding for message attitude or for emotional tone [28]. A clear distinction between the two could improve comparability of sentiment ratings across studies. If relevant to the research questions, Twitter studies may want to code images and URLs. This added step can be time-consuming, but doing so could change the context or meaning of a tweet. For example, an image could help determine if a tweet should be coded as humor or sarcasm, while following a URL could help determine if the tweet is an advertisement. Failing to code images and URLs could result in missing significant content that could affect coding accuracy and skew results. Analyzing Twitter user profiles could provide context for tweets. The same pro–e-cigarette tweet could hold very different significance for a study’s results if it is from a vape shop versus an adolescent. However, the limitations of analyzing user profiles should be recognized and steps taken to improve accuracy. It should be noted that user profiles may be misleading (eg, a tobacco industry representative posing as an unaffiliated citizen) or profiles of bots. Bots should be identified and potentially removed to reduce bias and improve the quality of data [57]. We cannot assume that the analysis is describing individuals but must instead consider the poster to be an “account” rather than a person. A limitation of Twitter data is that it does not provide much information on the effects of tweets on behavior. For example, tweeting about tobacco use does not necessarily mean that the person tweeting uses tobacco [40]. Additionally, an individual may tweet about a quit smoking campaign with positive sentiment, but that does not mean that the campaign has influenced his or her smoking behavior. Follow-up studies with the individuals tweeting about the campaign would need to be conducted. None of the studies in the review included follow-up survey studies with individuals from their Twitter sample. Although recruitment of individuals through Twitter may come with its own set of obstacles, this could be an avenue for future Twitter research. Following specific individuals over time could allow for the analysis of changes in message content.

Review Limitations

There are several limitations of this systematic review. First, the results are limited to the databases and search keywords selected, which could have resulted in incomplete retrieval of identified research. Second, this review is limited to its inclusion criteria and the decision rules of the single, independent coder who selected the articles and extracted the data to be included in the review. Individual bias was limited by explicit inclusion, exclusion, and data extraction criteria. However, some studies that were included or excluded for this review may have varied given different inclusion criteria or a different coder. Finally, this review is limited to methodology of categorically coded tobacco Twitter data. Different methodological results and recommendations may have been made if the topic of Twitter data had been different (eg, marijuana) or if the review had focused on different outcomes (eg, popularity of tweets or diffusion of tweets).

Conclusions

Categorically coded Twitter research can be used for certain insights that other survey research does not provide: emerging issues, popular content in real time, changes over time, how tobacco companies and pro-vaping advocacy groups use social media to increase message exposure in the population (eg, youth who might otherwise be protected from tobacco marketing), how tobacco control policies and campaigns can most effectively use social media, arguments by groups that may be incorporated into media message design, and quick reactions to antitobacco media campaigns and regulations. There are several approaches that researchers are taking to this end, each having its own set of strengths and weaknesses. Standards for data collection and coding should be developed to more easily compare and replicate tobacco-related Twitter results. Additional recommendations, dependent on one’s research goals, include the following: sample Twitter’s databases multiple times, make a distinction between message attitude and emotional tone for sentiment, code images and URLs, analyze user profiles, and identify and remove bots.
  28 in total

1.  Twitter=quitter? An analysis of Twitter quit smoking social networks.

Authors:  Judith J Prochaska; Cornelia Pechmann; Romina Kim; James M Leonhardt
Journal:  Tob Control       Date:  2011-07-05       Impact factor: 7.552

2.  Electronic Cigarettes on Twitter - Spreading the Appeal of Flavors.

Authors:  Kar-Hai Chu; Jennifer B Unger; Tess Boley Cruz; Daniel W Soto
Journal:  Tob Regul Sci       Date:  2015-03-01

3.  Are you Scared Yet?: Evaluating Fear Appeal Messages in Tweets about the Tips Campaign.

Authors:  Sherry L Emery; Glen Szczypka; Eulàlia Puig Abril; Yoonsang Kim; Lisa Vera
Journal:  J Commun       Date:  2014-04

4.  Vape, quit, tweet? Electronic cigarettes and smoking cessation on Twitter.

Authors:  Jan van der Tempel; Aliya Noormohamed; Robert Schwartz; Cameron Norman; Muhannad Malas; Laurie Zawertailo
Journal:  Int J Public Health       Date:  2016-02-03       Impact factor: 3.380

5.  Communicating genetics and smoking through social media: are we there yet?

Authors:  Sylviane de Viron; L Suzanne Suggs; Angela Brand; Herman Van Oyen
Journal:  J Med Internet Res       Date:  2013-09-09       Impact factor: 5.428

6.  Tweeting for and against public health policy: response to the Chicago Department of Public Health's electronic cigarette Twitter campaign.

Authors:  Jenine K Harris; Sarah Moreland-Russell; Bechara Choucair; Raed Mansour; Mackenzie Staub; Kendall Simmons
Journal:  J Med Internet Res       Date:  2014-10-16       Impact factor: 5.428

7.  Vaporous Marketing: Uncovering Pervasive Electronic Cigarette Advertisements on Twitter.

Authors:  Eric M Clark; Chris A Jones; Jake Ryland Williams; Allison N Kurti; Mitchell Craig Norotsky; Christopher M Danforth; Peter Sheridan Dodds
Journal:  PLoS One       Date:  2016-07-13       Impact factor: 3.240

8.  Using Twitter Data to Gain Insights into E-cigarette Marketing and Locations of Use: An Infoveillance Study.

Authors:  Annice E Kim; Timothy Hopper; Sean Simpson; James Nonnemaker; Alicea J Lieberman; Heather Hansen; Jamie Guillory; Lauren Porter
Journal:  J Med Internet Res       Date:  2015-11-06       Impact factor: 5.428

9.  A cross-sectional examination of marketing of electronic cigarettes on Twitter.

Authors:  Jidong Huang; Rachel Kornfield; Glen Szczypka; Sherry L Emery
Journal:  Tob Control       Date:  2014-07       Impact factor: 7.552

10.  Social Listening: A Content Analysis of E-Cigarette Discussions on Twitter.

Authors:  Heather Cole-Lewis; Jillian Pugatch; Amy Sanders; Arun Varghese; Susana Posada; Christopher Yun; Mary Schwarz; Erik Augustson
Journal:  J Med Internet Res       Date:  2015-10-27       Impact factor: 5.428

View more
  20 in total

1.  Toward Real-Time Infoveillance of Twitter Health Messages.

Authors:  Jason B Colditz; Kar-Hai Chu; Sherry L Emery; Chandler R Larkin; A Everette James; Joel Welling; Brian A Primack
Journal:  Am J Public Health       Date:  2018-06-21       Impact factor: 9.308

2.  I wake up and hit the JUUL: Analyzing Twitter for JUUL nicotine effects and dependence.

Authors:  Jaime E Sidani; Jason B Colditz; Erica L Barrett; Ariel Shensa; Kar-Hai Chu; A Everette James; Brian A Primack
Journal:  Drug Alcohol Depend       Date:  2019-08-30       Impact factor: 4.492

3.  Talking about tobacco on Twitter is associated with tobacco product use.

Authors:  Jennifer B Unger; Robert Urman; Tess Boley Cruz; Anuja Majmundar; Jessica Barrington-Trimis; Mary Ann Pentz; Rob McConnell
Journal:  Prev Med       Date:  2018-06-10       Impact factor: 4.018

4.  Expressed Symptoms and Attitudes Toward Using Twitter for Health Care Engagement Among Patients With Lupus on Social Media: Protocol for a Mixed Methods Study.

Authors:  Alden Bunyan; Swamy Venuturupalli; Katja Reuter
Journal:  JMIR Res Protoc       Date:  2021-05-06

Review 5.  Methods to Establish Race or Ethnicity of Twitter Users: Scoping Review.

Authors:  Su Golder; Robin Stevens; Karen O'Connor; Richard James; Graciela Gonzalez-Hernandez
Journal:  J Med Internet Res       Date:  2022-04-29       Impact factor: 7.076

6.  User Perceptions of Different Electronic Cigarette Flavors on Social Media: Observational Study.

Authors:  Xinyi Lu; Long Chen; Jianbo Yuan; Joyce Luo; Jiebo Luo; Zidian Xie; Dongmei Li
Journal:  J Med Internet Res       Date:  2020-06-24       Impact factor: 5.428

7.  Vaping associated with healthy food words: A content analysis of Twitter.

Authors:  Tatiana Basáñez; Anuja Majmundar; Tess Boley Cruz; Jennifer B Unger
Journal:  Addict Behav Rep       Date:  2018-09-24

8.  Perceived Attitudes About Substance Use in Anonymous Social Media Posts Near College Campuses: Observational Study.

Authors:  Alexis S Hammond; Michael J Paul; Joseph Hobelmann; Animesh R Koratana; Mark Dredze; Margaret S Chisolm
Journal:  JMIR Ment Health       Date:  2018-08-02

9.  Combining Crowd-Sourcing and Automated Content Methods to Improve Estimates of Overall Media Coverage: Theme Mentions in E-cigarette and Other Tobacco Coverage.

Authors:  Laura A Gibson; Leeann Siegel; Elissa Kranzler; Allyson Volinsky; Matthew B O'Donnell; Sharon Williams; Qinghua Yang; Yoonsang Kim; Steven Binns; Hy Tran; Veronica Maidel Epstein; Timothy Leffel; Michelle Jeong; Jiaying Liu; Stella Lee; Sherry Emery; Robert C Hornik
Journal:  J Health Commun       Date:  2019-11-13

10.  Eating Disorder Awareness Campaigns: Thematic and Quantitative Analysis Using Twitter.

Authors:  Iranzu Viguria; Miguel Angel Alvarez-Mon; Maria Llavero-Valero; Angel Asunsolo Del Barco; Felipe Ortuño; Melchor Alvarez-Mon
Journal:  J Med Internet Res       Date:  2020-07-14       Impact factor: 5.428

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.