Literature DB >> 36212768

Is academic writing becoming more positive? A large-scale diachronic case study of Science research articles across 25 years.

Zhou-Min Yuan1, Mingxin Yao2.   

Abstract

Academic writing is developing to be more positive. This linguistic positivity bias is confirmed in academic writing across disciplines and genres. The current research adopted sentiment analysis and examined the diachronic change in linguistic positivity in the full texts of 2,556 research articles published in Science in 25 years. The results showed that academic writing in research articles in the journal Science has become significantly more positive in the past 25 years. The findings of this study confirm linguistic positivity bias in academic writing based on empirical data from Science. Reasons for the increasingly positive language use in science articles might include the popularization of science, the growing number of researchers, and the difficulty of publishing in high-impact journals. Finally, this study discussed the implications of our findings for researchers, editors, and peer reviewers. © Akadémiai Kiadó, Budapest, Hungary 2022, Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Entities:  

Keywords:  Academic writing; Linguistic positivity bias; Research articles; Science; Sentiment analysis

Year:  2022        PMID: 36212768      PMCID: PMC9526210          DOI: 10.1007/s11192-022-04515-2

Source DB:  PubMed          Journal:  Scientometrics        ISSN: 0138-9130            Impact factor:   3.801


Introduction

Linguistic positivity bias (Augustine et al., 2011; Rozin et al., 2010) has received wide attention from various disciplines, such as psychology (Augustine et al., 2011; Rozin et al., 2010) and data science (Dodds et al., 2015; Garcia et al., 2012). Existing research reveals a general tendency to use more positive words than negative words in human communication. (Rozin et al., 2010; Augustine, 2011; Garcia et al., 2012; Dodds et al., 2015). For example, Dodds et al. (2015) observe a universal positivity bias in 100,000 words spread across 24 corpora in 10 languages with diverse origins. Augustine et al. (2011) find evidence for the human tendency to use positive words more often than negative words in written and spoken English. Linguistic positivity bias is likely to be universal across languages and registers. Linguistic positivity bias has also been identified in recent studies on academic writing, which find an upward trend in linguistic positivity in academic writing (Cao et al., 2020; Holtz et al., 2017; Lerchenmueller et al., 2019; Vinkers et al., 2015; Weidmann et al., 2018; Wen & Lei, 2021). Evidence in support of this claim is unearthed in academic articles across disciplines, including medical and life science (Cao et al., 2020; Lerchenmueller et al., 2019; Vinkers et al., 2015; Wen & Lei, 2021), political science (Weidmann et al., 2018), and cross-cultural psychology (Holtz et al., 2017). Vinkers et al. (2015) found significant growth in the frequency of positive words used in all scientific abstracts in PubMed from 1974 to 2014. Following Vinkers et al. (2015), Weidmann et al. (2018) also observe a marked increase in the use of positive words in abstracts in political science over time. Lerchenmueller et al. (2019) unveil the role of gender differences in the positive presentation of academic writing by examining titles and abstracts from a large dataset of academic articles in clinical research and life science in PubMed from 2002 to 2007. To date, these studies are largely limited to the examination of a linguistic positivity bias in the abstracts based on a small list of predefined positive/negative words. To overcome the limitations in previous studies, studies on linguistic positivity bias in academic writing also make various methodological modifications. First, full texts in academic journals are taken into account to validate the findings based on abstracts. For example, Holtz et al. (2017) find a general upward trend in the use of positive words based on the exploration of a linguistic positivity bias in the full texts of research articles from cross-cultural psychology. Cao et al. (2020) also find an increasingly positive trend in terms of linguistic positivity based on the examination of how the frequencies of positive and negative words change over time in both abstracts and full texts in journals from PubMed. Second, the latest studies resort to larger dictionaries and lexicons to tackle the limitation of the small list of positive and negative words (Vinkers et al., 2015), including studies by Holtz et al. (2017), Bordignon et al. (2021), and Wen and Lei (2021). In these studies, researchers adopt self-created dictionaries (Holtz et al., 2017), expanded list of positive/negative words (Bordignon et al., 2021), or sentiment analysis with large lexicons in R (Wen & Lei, 2021) to triangulate the results based on Vinkers et al.’s (2015) small list of positive and negative words. Third, regarding the limitation of findings generalised from one discipline, Bordignon et al. (2021) compare abstracts across disciplines between a pre-pandemic corpus and a corpus of preprints issued in response to the COVID-19 pandemic, and discover an increase of positive words and a slight decrease of negative words. The results are encouraging because in these studies, a growing trend of positive language is observed based on mixed methods, thus further confirming the linguistic positivity bias in academic writing. While existing studies provide significant insights into linguistic positivity bias in academic writing, the findings are still limited in the following three aspects. First, all of the exiting studies, except for Bordignon et al. (2021), investigated linguistic positivity bias in academic writing in individual disciplines. Little study has researched linguistic positivity bias in academic writing across disciplines. Second, most findings in the existing research are based on abstracts, the generalisability of which is open to doubt since abstracts may fail to fully and accurately reflect the main body of research articles (Pitkin and Branagan, 1998, Pitkin et al., 1999). Third, among the small number of studies that examined linguistic positivity bias in full texts (Cao et al., 2020 and Holtz et al., 2017), findings were generated based on a small list of predefined positive/negative words (Vinkers et al., 2015), which may not be sufficiently robust to detect sentiment polarity in academic writing from various disciplines (Wen & Lei, 2021: 4). To address these limitations in the literature, the current study intends to examine linguistic positivity bias in academic writing in the following ways. First, we attempt to investigate linguistic positivity bias in research articles in Science. As the leading outlet and the core journal of scientific discovery, Science is a multidisciplinary journal (Glänzel & Schubert, 2003; Glänzel et al., 1999) that welcomes research articles from all fields of science and any other source (https://www.science.org/content/page/mission-and-scope), and is thus an ideal sample to represent academic writing across disciplines. Second, we intend to explore linguistic positivity bias in the full texts of research articles published in Science for more in-depth findings. Finally, we adopt sentiment analysis with large lexicons of sentiment words (following Wen & Lei, 2021) in our data analysis for more robust and accurate results. Based on these considerations, this study addresses how the sentiment in the full texts of Science has changed in the past 25 years. Finally, we intend to discuss the possible reasons for the development of sentiment of the full texts in Science in the past 25 years.

Methods

This section presents the data collection and data analysis that were employed in the current research.

Data Collection

According to the categories of manuscripts given by Science (https://www.science.org/content/page/science-information-authors), the journal accepts and publishes the following types of manuscripts: research articles, reports, reviews, and commentaries. In our study, we only focused on the category of research articles for the following reasons. First, analysis on this single category eliminates the influence of such factors as genre differences between research articles and other categories. Second, concerning that the categories may change over time, we chose the category of research articles since it has remained consistent over the past 25 years. Finally, the reason why we did not include reports in our study is that research articles and reports may vary in terms of academic significance and paper length. Specifically, the former presents research papers with major advances containing up to 4500 words while the latter only accepts important new research results of broad significance and is limited to 2500 words. In the current research study, we developed a script in Python for data collection, based on which a diachronic corpus for academic writing was constructed. The data were retrieved from the journal Science through access provided by Nanjing University Library. During data collection, the crawl delay was set to 10, conforming to the Robots Exclusion Standard announced by the Journal website.1 The collected data were stored in the format of Excel files to form a corpus with 25 years of research articles (full texts) published in Science. The corpus consisted of 2,556 research articles dating from January 1997 to August 2021,2 with a total of 10,915,515 words (see Table 1 for the descriptive statistics), representing academic writing samples for a wide variety of fields and disciplines. The full texts of these research articles were published across a span of 25 years, sufficient time to explore the diachronic change of linguistic positivity in academic writing. In addition, the investigation on the full texts allows us to gain a holistic understanding of academic writing, hence yielding more reliable and generalisable results than those from analysing only abstracts.
Table 1

Descriptive statistics of the corpus used in the study

YearNumber of ArticlesNumber of Words in Full TextsMean Word CountNumber of Sentences Per YearMean Sentence Count Per Article
199735148,0554,2308,712248
199845174,0713,86811,561256
199944169,4133,85010,800245
200059227,9023,86215,980270
200167252,7383,77215,894237
200267248,0883,70215,471230
200360226,6243,77714,663244
200460228,1063,80114,071234
200572266,9533,70716,793233
200664236,7203,69814,455225
200758226,3353,90214,653252
200857211,8283,71613,318233
200969275,5203,99319,008275
201067268,9914,01419,143285
201175292,8173,90420,723276
201254213,2153,94814,809274
201390386,5484,29426,018289
2014126551,8054,37937,231295
2015114507,9894,45633,636295
2016149669,6594,49445,393304
2017172780,3694,53752,982308
2018192875,8194,56160,291314
20192581,179,6684,57286,932336
20202781,256,4434,51993,645336
20212241,039,8394,64279,431354
Descriptive statistics of the corpus used in the study

Data analysis

Sentiment analysis

To address the research question, we conducted sentiment analysis, a method that studies the positive or negative evaluations, attitudes, and views expressed in a text (Liu & Lei, 2018; Mäntylä et al., 2018; Serrano-Guerreroet al., 2015; Taboada, 2016; Wen & Lei, 2021). To date, two main approaches have been commonly adopted in sentiment analysis: machine learning and lexicon-based approaches (Taboada, 2016; Mukhtar, 2018; Van Houtan et al., 2020; Wen & Lei, 2021). The former approach runs on a classifier trained for determining the polarity of texts (Taboada, 2016). However, this approach is limited to the specific field or genre of the training dataset that the classifier is trained for (Wen & Lei, 2021: 7). The lexicon-based approach, on the other hand, is based on lexicons or dictionaries containing a large set of sentiment words and their polarities (Taboada, 2016; Wen & Lei, 2021). This approach, although less accurate, is not subject to a particular genre or domain of the trained texts and, therefore, can efficiently handle data from different domains (Mukhtar, 2018: 2182). Before sentiment analysis, we preprocessed the raw texts in our corpus by removing all the non-English symbols such as β or Ӓ, which may also be used in equations, etc., to ensure that our analysis is not affected by such special symbols. To do so, we coded a regular expression in python, which deletes the non-English symbols in our data. In the current study, we employ the lexicon-based approach in sentiment analysis for the following reasons. First, this method is proven to be robust across different domains without changing the dictionaries (Taboada et al., 2011: 9). Second, this approach allows us to compare our findings with previous studies using the same approach, such as Wen and Lei (2021). In detail, we coded an R script to run sentiment analyses on each research article (full text). Two packages embedded in R are used separately for performing sentiment analysis, namely, Syuzhet (Jockers, 2017) and Sentimentr (Rinker, 2018). In the first sentiment analysis (SA1), we resort to Syuzhet (Jockers, 2017), a popular R package widely applied in studies with sentiment analysis (Bradley & James, 2019; Jensen & Bang, 2017; Liu & Lei, 2018; Vergeer, 2020; Wen & Lei, 2021). However, Syuzhet (Jockers, 2017) was found to be error-prone due to the lack of valence shifters (i.e., negators, intensifiers, or downtoners) in a sentence (See Rinker, 2018). Later, Rinker (2018) released a modified R package, Sentimentr (Rinker, 2018), based on the weakness of Syuzhet (Jockers, 2017). Therefore, the second sentiment analysis (SA2) is carried out based on Sentimentr (Rinker, 2018). In addition, with two packages, we are able to triangulate the results by comparing those of SA1 and SA2. We also ran sentiment analysis on multiple lexicons in Syuzhet (Jockers, 2017) and Rinker (2018) to further triangulate the results. Specifically, in Syuzhet (Jockers, 2017), we opted for the Jockers sentiment lexicon (Jockers, 2017) and the NRC sentiment lexicon (Mohammad & Turney, 2010). In Sentimentr (Rinker, 2018), we used three lexicons, including the previous two lexicons and one additional lexicon, i.e., the SenticNet lexicon (Cambria et al., 2016). The reasons for using these lexicons are as follows. First, these lexicons are proven to be highly robust (Mohammad, 2010) and reliable (Wen & Lei, 2021) in terms of sentiment analysis. Second, embedded in the R packages, these lexicons are free and open for access. The sentiment analysis procedure based on the above lexicons in Syuzhet (Jockers, 2017) and Sentimentr (Rinker, 2018) follows several steps. It should be noted that all of the algorithms used for sentiment analysis in this study are based on sentences. To calculate the sentiment scores for one article, the algorithms in the R packages first divide each research article (full text) into individual sentences. Next, they produce a raw sentiment score for each sentence in the article. Finally, a composite sentiment score for each article is calculated by adding up the raw scores of all the sentences in the article. However, the results produced by different lexicons are not comparable because the sentiment words included in these lexicons were tagged on different scales and intervals (Wen & Lei, 2021: 9). Therefore, standardization of the raw sentiment scores is necessary after the analysis. We followed Lennox et al. (2020) and Wen and Lei (2021) in standardizing the raw sentiment scores by calculating the mean, i.e., μ (sentiments) and the standard deviation, i.e., σ (sentiments) of the raw scores of all the research articles, and finally the standardized sentiment score for each article based on Lennox et al.’s method (2020), as displayed in Formula 1. Finally, to compare the sentiment scores across time on a yearly basis, we calculated the yearly means of the standardized sentiment scores. Formula 1: In addition, it should be noted that sentiment analysis with these two packages is not affected by factors such as the prevalence of positive words and sentence length (Wen & Lei, 2021), due to a larger proportion of negative words in the lexicons and the design of the algorithms. However, due to the space limit and also the scope of the present paper, we do not specify the technical details for our instruments. For more technical details, please see Rinker (2019) as a consultation.

Statistical analysis

In terms of statistical analysis, we first conducted simple linear regression (Lei & Wen, 2020; Lei & Yan, 2016; Lei & Zhang, 2018; Wen & Lei, 2021) to examine the diachronic development of research articles in Science in terms of sentiment scores. Specifically, we performed simple linear regression on all five result samples from the five lexicons, two of which were based on Syuzhet (Jockers, 2017) and the other three based on Sentimentr (Rinker, 2018). In all the analyses of simple linear regression, we examined the developmental trajectory of the sentiment scores with the year as the independent variable and the standardized sentiment score of the full text of each research article as the dependent variable. In addition, to further compare the results of sentiment analysis based on different packages and lexicons, we also performed Pearson’s product-moment correlation analyses (Wen & Lei, 2021) to examine whether the five result samples are positively or negatively correlated with statistical significance.

Results

This section first reports the distribution of sentiment in the full texts across 25 years and then the results of statistical analysis, which may shed light on the trend of linguistic positivity bias.

Distribution of sentiment across 25 years

We performed sentiment analyses based on five lexicons in two R packages, i.e., Jockers sentiment lexicon and NRC sentiment lexicon in Syuzhet (Jockers, 2017), as well as Jockers sentiment lexicon, NRC sentiment lexicon, and SenticNet sentiment lexicon in Sentimentr (Rinker, 2018). In this step, our data analysis generated five result samples for further analysis. To compare the sentiment scores across time on a yearly basis, we calculated the yearly means of the standardized sentiment scores from January 1997 to August 2021, as displayed in Table 2.
Table 2

Distribution of sentiment in the full texts across 25 years

YearSyuzhet (Jockers, 2017)Sentimentr (Rinker, 2018)
Jockers sentiment lexiconNRC sentiment lexiconJockers sentiment lexiconNRC sentiment lexiconSenticNet sentiment lexicon
19979.5213319.52753310.08001410.66195211.987791
19989.3965319.4065499.87016610.31590811.814906
19997.5461217.5786377.9997148.218369.868191
20008.5714318.6089579.1732259.52673410.960807
20018.3832778.4339248.9988619.4200910.892345
20027.8637077.8985258.5111718.91734610.222025
20037.7779687.8132378.3504948.48809410.203304
20048.4319918.4626088.9790329.34479110.948697
20057.4469017.4392317.9246718.1100299.693471
20067.5800087.587668.2275868.48996410.051596
20078.7355938.7758769.381169.68249111.204487
20088.0544968.1057458.5381128.9414610.547337
200910.10857510.15324510.77765911.06963412.64925
201010.25788210.27727211.07193411.16650712.782353
201110.29561910.31827911.03311311.37683112.711169
201210.10142510.11582610.78634211.12461112.645313
20139.9042279.94726910.57472710.81586312.312222
201410.46643310.5024211.11191511.36812212.987398
201511.02050611.05024611.77551811.97722113.53401
201611.33075611.35005512.02389812.19256313.852181
201711.59167111.61740212.24530712.45797114.068735
201811.22581811.24614911.84653512.01621713.699407
201910.60349910.62028311.20295111.35516913.117856
202010.81210.8370811.45439711.64857213.344973
202110.41397710.45118210.98813111.14270912.929841
Distribution of sentiment in the full texts across 25 years From eyeballing the data, we derived the hypothesis that the sentiment scores in the full texts went through a general increase from 1997 to 2021. However, since the statistical evaluation goes beyond what eyeballing the table can do (Hilpert & Gries, 2009: 390), we must rely on further statistics for precise insights into the diachronic change of sentiment scores in the data, which is presented in the next section.

Trends of linguistic positivity based on sentiment analyses

In this section, we first report the statistical result of SA1 before that of SA2 because they are based on distinct packages.

Results of the SA1

In SA1, we ran sentiment analysis with two lexicons in Syuzhet (Jockers, 2017), namely, the Jockers sentiment lexicon and the NRC sentiment lexicon. Table 3 summarizes the descriptive statistics of the standardized sentiment scores based on each lexicon. Figure 1 demonstrates the yearly means of the standardized sentiment scores based on SA1 from January 1997 to August 2021.
Table 3

Descriptive statistics of standardized sentiment scores in SA1

LexiconMeanaStandard deviationaMaximumMinimum
Jockers10.7430971.82587338.412861− 17.274664
NRC9.497671.3404838.418873− 17.316931

amean and standard deviation by year and full text

Fig. 1

Diachronic trajectory of linguistic positivity based on SA1

Descriptive statistics of standardized sentiment scores in SA1 amean and standard deviation by year and full text Diachronic trajectory of linguistic positivity based on SA1 The results of simple linear regression on SA1 suggested a significant increase in sentiment in the full texts, indicating an upward developmental trend of linguistic positivity in the last 25 years (Jockers: F(1,23) = 34.15, p = 5.912e-06, multiple R2 = 0.5975, adjusted R2 = 0.58; NRC: F(1,23) = 34.33, p = 5.694e-06, multiple R2 = 0.5988, adjusted R2 = 0.5814;). Table 4 presents the detailed statistics of the model.
Table 4

Detailed statistics of simple linear regression for SA1

ModelVariableEstimateStandard errort-valuep-value
Jockers(Intercept)− 273.3480548.40457− 5.6479.51e− 06 ***
Year0.140790.024095.8435.91e− 06 ***
NRC(Intercept)− 273.4247948.29356− 5.6629.18e− 06 ***
Year0.140840.024045.8595.69e− 06 ***

*p < 0.05; **p < 0.01; ***p < 0.001

Detailed statistics of simple linear regression for SA1 *p < 0.05; **p < 0.01; ***p < 0.001

Results of the SA2

In SA2, we analysed the sentiments in the full texts with Sentimentr (Rinker, 2018) and three lexicons in the package, i.e., the Jockers sentiment lexicon (Jockers, 2017), the NRC sentiment lexicon (Mohammad & Turney, 2010), and the SenticNet lexicon (Cambria et al., 2016). Table 5 demonstrates the descriptive statistics of the standardized sentiment scores with SA2. Figure 2 displays the yearly means of the standardized sentiment scores of the full texts across 25 years.
Table 5

Descriptive statistics of the standardized sentiment scores in SA2

LexiconMeanaStandard deviationaMaximumMinimum
Jockers10.4922731.62829939.64096− 18.735906
NRC10.6173421.68951940.348968− 19.887274
SenticNet10.7294281.83554340.884626− 17.719674

amean and standard deviation by year and full text

Fig. 2

Diachronic trajectory of linguistic positivity based on SA2

Descriptive statistics of the standardized sentiment scores in SA2 amean and standard deviation by year and full text Diachronic trajectory of linguistic positivity based on SA2 In the statistical analysis for SA2, simple linear regression revealed an upward trend in the development of linguistic positivity because the sentiments have significantly increased in the full texts in the past 25 years (Jockers: F(1,23) = 34.26, p = 5.773e− 06, multiple R2 = 0.5983, adjusted R2 = 0.5809; NRC: F(1,23) = 27.38, p = 2.625e− 05, multiple R2 = 0.5435, adjusted R2 = 0.5236; SenticNet: F(1,23) = 34.63, p = 5.346e− 06, multiple R2 = 0.6009, adjusted R2 = 0.5836). Table 6 demonstrates the results of the simple linear regression model.
Table 6

Detailed statistics of simple linear regression for SA2

ModelVariableEstimateStandard errort-valuep-value
Jockers(Intercept)− 283.9781950.24481− 5.6529.40e− 06 ***
Year0.146390.025015.8535.77e− 06 ***
NRC(Intercept)− 262.8676352.22310− 5.0344.29e− 05 ***
Year0.146390.025015.8535.77e− 06 ***
SenticNet(Intercept)− 281.3278949.83593− 5.6459.56e− 06 ***
Year0.145990.024815.8855.35e− 06 ***

*p < 0.05; **p < 0.01; ***p < 0.001

Detailed statistics of simple linear regression for SA2 *p < 0.05; **p < 0.01; ***p < 0.001 Generally, the statistical results of SA2 are in line with those of SA1, thus offering further triangulated evidence for the increasing linguistic positivity in academic writing during the 25 years evaluated.

Results of the correlation test

Finally, we performed Pearson’s correlation test (Wen & Lei, 2021) on the five samples of standardized sentiment scores in SA1 and SA2 to test the reliability of sentiment analysis. Table 7 shows the interrelation of the standardized sentiment scores in SA1 and SA2. According to the results, all the sentiment measures are highly significantly correlated (r = 0.9999409 > 0.9, p = 2.2e− 16 < 0.001), providing further triangulated evidence for the reliability of our sentiment analysis.
Table 7

Pearson’s correlation test for standardized sentiment scores in SA1 and SA2

Lexicon12345
1. Jockers (SA1)
2. NRC (SA1)0.9999409***
3. Jockers (SA2)0.9986053***0.9986208***
4. NRC (SA2)0.9957922***0.9958604***0.9966371***
5. SenticNet (SA2)0.9990544***0.9991685***0.9981632***0.9959328***

*p < 0.05; **p < 0.01; ***p < 0.001

Pearson’s correlation test for standardized sentiment scores in SA1 and SA2 *p < 0.05; **p < 0.01; ***p < 0.001

Fluctuations in the sentiment scores

The current study investigated linguistic positivity bias in academic writing based on a diachronic corpus of academic research articles published between January 1997 and August 2021 in Science. To date, this research is most likely the first study that employs sentiment analysis with large lexicons to examine linguistic positivity bias in the full texts, rather than abstracts, of academic writing from a diachronic perspective. The findings of this study revealed a generally increasingly positive trend in academic writing in Science over the past 25 years. Our findings were in line with those in Cao et al. (2020) and Wen and Lei (2021) with evidence from Science. Our results also fill the gap that no previous study has examined linguistic positivity bias by employing sentiment analysis with large lexicons on the full texts of research articles from a diachronic perspective. However, some fluctuations were observed in the data. Roughly between 1999 and 2008 there were some fluctuations in the sentiment scores, which is followed by a general increase in the scores approximately from 2009 to 2018. We find it necessary to report and discuss why there were such fluctuations in our data. To do so, we investigated the linguistic nature of our corpus with the help of a list of predefined positive and negative words proposed by Vinkers et al. (2015), as displayed in Table 8. Specifically, we coded a python script to calculate the frequency of these positive and negative words in the corpus. The above-mentioned fluctuations between 1999 and 2008 were also observed in the frequency of Vinkers et al’s (2015) positive and negative words in our corpus, as displayed in Fig. 3. From 2009 to 2018 there was a general increase in the sentiment scores, which was also in line with the frequency of Vinkers et al’s (2015) positive and negative words between 2009 and 2018, as shown in Fig. 4. To be specific, there appeared to be more increasingly more positive words than negative words from 2009 to 2018.
Table 8

List of predefined positive and negative words (Vinkers et al., 2015)

CategoryWords
Positiveamazing, assuring, astonishing, bright, creative, encouraging, enormous, excellent, favourable, groundbreaking, hopeful, innovative, inspiring, inventive, novel, phenomenal, prominent, promising, reassuring, remarkable, robust, spectacular, supportive, unique, unprecedented
Negativedetrimental, disappointing, disconcerting, discouraging, disheartening, disturbing, frustrating, futile, hopeless, impossible, inadequate, ineffective, insignificant, insufficient, irrelevant, mediocre, pessimistic, substandard, unacceptable, unpromising, unsatisfactory, unsatisfying, useless, weak, worrisome
Fig. 3

Frequency of Vinkers et al’s (2015) positive and negative words between 1999 and 2008

Fig. 4

Frequency of Vinkers et al’s (2015) positive and negative words between 2009 and 2018

List of predefined positive and negative words (Vinkers et al., 2015) Frequency of Vinkers et al’s (2015) positive and negative words between 1999 and 2008 Frequency of Vinkers et al’s (2015) positive and negative words between 2009 and 2018 The overall distribution of Vinkers et al’s (2015) positive and negative words between 1997 and 2021 is illustrated in Fig. 5. From Fig. 5, we are able to see a general increase in the frequency of positive words, with less fluctuations observed. This result might further reinforce our arguments based on the results of the previous sentiment analysis that academic writing in Science research articles has become increasingly positive.
Fig. 5

Overall distribution of Vinkers et al’s (2015) positive and negative words between 1997 and 2021

Overall distribution of Vinkers et al’s (2015) positive and negative words between 1997 and 2021 However, we have to acknowledge the limitations of Vinkers et al’s (2015) list of positive and negative words since they are small in number, and therefore could not entirely represent the fluctuations in our sentiment scores. Therefore, such an analysis merely provide an aspect of the linguistic variations in the corpus. In other words, the fluctuations in our sentiment scores could possibly be a result of the inner linguistic variations in the corpus, with Vinkers et al’s (2015) list of positive and negative words serving as a case in point.

Discussion

In this section, we discuss the potential reasons for the use of increasingly positive language in academic writings. First, the popularization of science (Bell & Turney, 2014; Pilkington, 2016) might be one reason why academic writing has developed to be more positive. The popularization of science is a long-standing tradition that includes a variety of practices in making scientific information more accessible to general and nonexpert audiences (Bucchi & Trench, 2014). As a result, both publishers and researchers are keen to promote scientific advancements and research brands with the public (Bell & Turney, 2014; Pilkington, 2016). By doing so, researchers tend to adopt narratives that shape a positive image of themselves as creative thinkers when describing their discoveries (Pilkington, 2016). In relation to Science, the journal seeks to not only advance scientific understanding but also publish papers that merit recognition by the wider scientific community and the general public (https://www.science.org/content/page/mission-and-scope). Therefore, it is possible that the popularization of science has played a vital role in the increasing linguistic positivity trend in research articles in Science over the past 25 years. Second, we agree with Wen and Lei (2021), Cao et al. (2020), and Vinkers et al. (2015) that the linguistic positivity bias is possibly influenced by competition in publications in the academic community. Specifically, positive language as a technique or strategy has been increasingly adopted in recent decades (Cao et al., 2020: 4), during which high-quality publications have gained such importance that they can influence various aspects of a researcher’s career (Nicolini & Nozza, 2008; Nosek et al., 2012; Wen & Lei, 2021), such as hiring, salary, promotion, tenure, and grant awards (Nosek et al., 2012: 616). At the same time, however, it has become increasingly more difficult to have one’s research published (Wen & Lei, 2021: 17) due to the high demand for publication (Nosek et al., 2012), the growing number of researchers (Lillis & Curry, 2013), the competitive process of publication selection (Millar et al., 2019), and the more thorough and critical editorial and peer review process in high-impact journals (Vinkers et al., 2015: 3). Consequently, researchers may adopt a more positive writing approach in research articles (Vinkers et al., 2015) to promote their research for publication purposes (Cao et al., 2020; Wen & Lei, 2021). Third, the increasing linguistic positivity bias in academic writing may also be attributed to positive outcome reporting bias (Dwan et al., 2013) or positive publication bias (Mlinarić et al., 2017). Recent studies found that studies with positive or statistically significant results have greater odds of being published (Dwan et al., 2013; Mlinarić et al., 2017), which contributes to a scientific culture that favors positive outcomes (Wen & Lei, 2021). As a result, researchers are more likely to report statistically positive or significant outcomes (Dwan et al., 2013) as a strategy to impress the audience (Wen & Lei, 2021), editors, and peer reviewers (Chiu et al., 2017). In this process of promoting their research (Cao et al., 2020; Millar et al., 2019), researchers may be inclined to use more positive writing, such as hyperbolic and/or subjective language, to glamorize and promote and/or exaggerate aspects of their research (Millar et al., 2019: 139). Researchers who have observed this phenomenon are now warning the scientific community about the risk that this practice may undermine the objectivity and interpretation of newly discovered scientific knowledge (Millar et al., 2019; Wen & Lei, 2021) as well as the trustworthiness of published research findings (Ioannidis, 2005). This study has implications for researchers, editors, and peer reviewers. On the one hand, positive language is useful in terms of selling the paper and promoting science to the general public; however, extensive use of positively subjective language could erode the accuracy of the information conveyed and, hence, result in doing a disservice to science (Millar et al., 2019). Therefore, researchers should adopt the right judgment of intention in academic writing (Millar et al., 2019) and a prudent use of promotional language (Cao et al., 2020) to preserve the integrity of the scientific findings. Wen and Lei (2021) also argue that researchers take responsibility for the language used in academic writing. The findings of this research and previous studies may also have some implications for editors and peer reviewers regarding the need for heightened vigilance for promotional language use (Millar et al., 2019) and more tolerance and even rewards for negative research results for the purpose of keeping science on track. (Nature Editorial, 2017). Empirically, our study features originality in the following fronts. First, we investigate the full texts of research articles on Science. Therefore, our findings based on the full texts may be more generalisable than those generated from abstracts. It may further validate and reinforce previous findings based on abstracts, whose generalisability might be questionable. Second, our study may better represent academic writing. Specifically, since Science publishes research articles from across disciplines, our findings could reveal the diachronic development of linguistic positivity in academic writing across disciplines instead of one or two disciplines. However, our study is also limited as follows. First, as a case study on Science, the findings on the full texts are limited to this journal. Future research may extend the size of the corpus by incorporating full texts from more scientific journals such as Nature and other Nature indexed journals. Second, our study revealed only a positive trend in the diachronic development of academic writing in science. However, it may fail to account for exactly how the use of positive language evolved over time. Future studies may approach the same issue with in-depth qualitative methods. Third, our study is limited in that we did not take into account that language use and hence sentiments may vary in different sections of a research article, which is a potentially interesting and relevant topic for future research. Finally, although our instruments exhibit robustness and high reliability, they are still limited in terms of accuracy. Future research may adopt tools with better accuracy in detecting sentiments in academic writing, such as more accurate algorithms or machine learning models.
  16 in total

1.  Can the accuracy of abstracts be improved by providing specific instructions? A randomized controlled trial.

Authors:  R M Pitkin; M A Branagan
Journal:  JAMA       Date:  1998-07-15       Impact factor: 56.272

2.  Sentiment analysis as a measure of conservation culture in scientific literature.

Authors:  Robert J Lennox; Diogo Veríssimo; William M Twardek; Colin R Davis; Ivan Jarić
Journal:  Conserv Biol       Date:  2019-10-01       Impact factor: 6.560

3.  Rewarding negative results keeps science on track.

Authors: 
Journal:  Nature       Date:  2017-11       Impact factor: 49.962

4.  A Positivity Bias in Written and Spoken English and Its Moderation by Personality and Gender.

Authors:  Adam A Augustine; Matthias R Mehl; Randy J Larsen
Journal:  Soc Psychol Personal Sci       Date:  2011-09

5.  Human language reveals a universal positivity bias.

Authors:  Peter Sheridan Dodds; Eric M Clark; Suma Desu; Morgan R Frank; Andrew J Reagan; Jake Ryland Williams; Lewis Mitchell; Kameron Decker Harris; Isabel M Kloumann; James P Bagrow; Karine Megerdoomian; Matthew T McMahon; Brian F Tivnan; Christopher M Danforth
Journal:  Proc Natl Acad Sci U S A       Date:  2015-02-09       Impact factor: 11.205

6.  Scientific Utopia: II. Restructuring Incentives and Practices to Promote Truth Over Publishability.

Authors:  Brian A Nosek; Jeffrey R Spies; Matt Motyl
Journal:  Perspect Psychol Sci       Date:  2012-11

7.  Over-promotion and caution in abstracts of preprints during the COVID-19 crisis.

Authors:  Frederique Bordignon; Liana Ermakova; Marianne Noel
Journal:  Learn Publ       Date:  2021-07-21

8.  Why most published research findings are false.

Authors:  John P A Ioannidis
Journal:  PLoS Med       Date:  2005-08-30       Impact factor: 11.613

Review 9.  Systematic review of the empirical evidence of study publication bias and outcome reporting bias - an updated review.

Authors:  Kerry Dwan; Carrol Gamble; Paula R Williamson; Jamie J Kirkham
Journal:  PLoS One       Date:  2013-07-05       Impact factor: 3.240

Review 10.  'Spin' in published biomedical literature: A methodological systematic review.

Authors:  Kellia Chiu; Quinn Grundy; Lisa Bero
Journal:  PLoS Biol       Date:  2017-09-11       Impact factor: 8.029

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.