Literature DB >> 33371013

Gender and ethnic differences in publication of BMJ letters to the editor: an observational study using machine learning.

Mohamad Zeina1, Alfred Balston1, Amitava Banerjee2, Katherine Woolf3.   

Abstract

OBJECTIVES: To analyse the relationship between first author's gender and ethnicity (estimated from first name and surname), and chance of publication of rapid responses in the British Medical Journal (BMJ). To analyse whether other features of the rapid response account for any gender or ethnic differences, including the presence of multiple authors, declaration of conflicts of interests, the presence of Twitter handle, word count, reading ease, spelling and grammatical mistakes, and the presence of references.
DESIGN: A retrospective observational study.
SETTING: Website of the BMJ (BMJ.com). PARTICIPANTS: Publicly available rapid responses submitted to BMJ.com between 1998 and 2018. MAIN OUTCOME MEASURES: Publication of a rapid response as a letter to the editor in the BMJ.
RESULTS: We analysed 113 265 rapid responses, of which 8415 were published as letters to the editor (7.4%). Statistically significant univariate correlations were found between odds of publication and first author estimated gender and ethnicity, multiple authors, declaration of conflicts of interest, the presence of Twitter handle, word count, reading ease, spelling and grammatical mistakes, and the presence of references. Multivariate analysis showed that first author estimated gender and ethnicity predicted publication after taking into account the other factors. Compared to white authors, black authors were 26% less likely to be published (OR: 0.74, CI: 0.57-0.96), Asian and Pacific Islander authors were 46% less likely to be published (OR: 0.54, CI: 0.49-0.59) and Hispanic authors were 49% less likely to be published (OR: 0.51, CI: 0.41-0.64). Female authors were 10% less likely to be published (OR: 0.90, CI: 0.85-0.96) than male authors.
CONCLUSION: Ethnic and gender differences in rapid response publication remained after accounting for a broad range of features, themselves all predictive of publication. This suggests that the reasons for the differences of these groups lies elsewhere. © Author(s) (or their employer(s)) 2020. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Keywords:  epidemiology; health economics; health informatics; health policy; medical education & training; statistics & research methods

Mesh:

Year:  2020        PMID: 33371013      PMCID: PMC7754629          DOI: 10.1136/bmjopen-2020-037269

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


This study uses corpus of publicly available data to analyse correlations between first author’s characteristics and chance of publication of letters to the editor in the British Medical Journal. Multivariate analysis allowed us to account for a range of other features of submitted letters. To the best of our knowledge, this is the largest ever analysis of a scientific corpus that looks at associations with publication rate. The nature of this data means that only associations can be inferred, and not causation. We highlight automated techniques that scientific journals can use to look for associations between ethnicity and gender in their own publication rates.

Introduction

Much has been written about the ‘attainment gap’ or ‘differential attainment’, the observation that many fields exhibit discrepancies in achievement based on personal attributes such as gender and ethnicity. In medicine, for example, students from black and minority ethnic (BME) groups achieve poorer marks and are more likely to fail, on average, than their white counterparts.1 As they progress in their careers, BME doctors also more often fail their specialty training exams,1 2 earn a lower average salary than others at the same level of seniority3 and are less likely to be awarded funding grants.4 There also remain discrepancies in the representation of women in medical leadership and faculty despite a long history of roughly equal proportions of male and female medical students.5 6 Another specific area where the effects of gender have been studied thoroughly is academic publishing. A survey of 1065 authors from different backgrounds found that women were under-represented in the scientific literature, along with certain ethnic minorities.7 In a group of high-impact medical journals, including the British Medical Journal (BMJ), the number of articles with female first authors has increased over time; however, the gender balance of last authors, who are typically senior researchers or heads of departments, has not followed this trend.8 A larger scale analysis of 1.8 million scholarly articles indexed in JSTOR found female authors are poorly represented in the prestigious first and last author positions in the majority of academic disciplines.9 The same result was observed in an analysis of 21 million articles indexed in Medline, which also found BME authors are similarly less likely to be in the last author position even when accounting for seniority.10 The cause of these gender and ethnic differences remains the subject of debate. Experimental studies have shown that identical submissions randomly assigned to have a male or female name are ranked differently depending on the gender of the applicant’s name, favouring men.11 12 For example, applications for a laboratory manager position that were assigned a male name were rated as being significantly more competent and hireable by faculty members compared with identical applications assigned a female name.11 Similarly, a study presented graduate students with a sample of abstracts from an international conference, where abstracts were randomly shown to have male or female authors. The study found that abstracts presented as having female authors were deemed to have lower ‘Scientific Quality’.12 Both papers found that the gender of the reviewer did not affect how applicants were rated, and concluded that pervasive gender stereotypes create a subtle but significant bias against women.11 12 Indeed, the bias was reduced when reviewers’ attitudes towards gender roles were taken into account, with higher support for gender equality being associated with higher ratings for female authors. Some argue that gender disparities arise from men and women choosing different career paths.13 For example, women may opt to prioritise flexibility or take time out of their career to have children. However, studies that incorporate these factors into multivariate statistical models fail to fully account for discrepancies in pay.14 15 One such study found a $14 581 yearly salary difference between male and female hospital physicians in the USA, which remained after accounting for differences in job satisfaction priorities between genders.16 The underlying causes of these differences are likely to be complex and multifactorial, but identifying and characterising disparities in new specific situations might hint at potential solutions. These may be broadly applicable, especially because the causative issues are likely to compound each other. For example, lower average pay for women and BME doctors may be partly due to lower chances of scientific publication, especially in a work environment where publication in the scientific literature is important for attaining certain senior academic and leadership positions. Though many studies mentioned here find group differences based on personal characteristics, they rarely have access to raw data from the journals that would be necessary to quantify publication rate. For example, a finding that women are under-represented in authorship of medical journal papers compared with in the medical workforce is not enough to draw conclusions about discrimination or bias, as it may be due to differences in priorities and the number of submissions sent. For a study to draw meaningful conclusions regarding discrepancies in acceptance rates, it must be able to quantify the percentage of submitted scientific works that are accepted, and this submission data is seldom released by scientific journals. Letters in the BMJ are derived from rapid responses, which are available online freely and in their entirety, and, therefore, they may provide a valuable perspective for looking at this issue. Moreover, publication of rapid responses is of importance since letters to the editor carry PubMed identifiers and thus discrepancies in their publication may have knock-on effects for jobs in academia where PubMed indexed publications play an important role in candidate selection. We aimed to compare the corpus of available rapid responses with published letters to the editor to look for correlations between ethnicity, gender and odds of publication.

Methods

Data acquisition and processing

An automated script was used to download every BMJ.com online rapid response between 25 April 1998 and 23 March 2018, as well as every letter to the editor that was published in the same timeframe. To minimise the impact on BMJ servers, webpage requests were only sent every 15 s, and each request explicitly stated a full name and contact email address of the researcher carrying out the automated data collection, so that they could easily be contacted if the BMJ wished this collection to stop. Furthermore, we only collected publicly available data that can be accessed without a login to a BMJ account. Once collected, every available field from the rapid response was extracted. This included: title, title of article being responded to, body of text, first author name, first author title, other authors, date of submission and the presence of Twitter handle. Further processing with software packages mentioned below allowed us to look at a richer set of features, including word count, the presence of references, number of references, Flesch reading ease (a measure of complexity of language, with a higher value meaning easier to read), number of spelling and grammatical mistakes, gender of first author, ethnicity of first author and the presence of multiple authors. The position of the author was extracted by looking for the presence of each of the words ‘Consultant’, ‘Professor’, ‘Senior’ and ‘Student’ in the self-declared occupation field of submitted rapid responses, for example, someone who had the word ‘Consultant’ anywhere in their occupation field was classed ‘Consultant’. We did not expect a linear relationship between publication and word count or Flesch reading ease, because the most successful letters are likely to be long enough to offer a meaningful insight into the topic, but not too long as to be unsuitable for the short letter to the editor format. Thus, we created two additional features from these, ‘Near ideal word count’ and ‘Near ideal Flesch reading ease’ to reflect whether a rapid response was within the 50% of rapid responses which are closest in word count and Flesch reading ease to the numbers which have historically been associated with higher rates of publication. Some rapid responses (528, or 0.46%) could not be collected automatically due to errors in their formatting which prohibited their automated collection. These rapid responses were omitted from analysis. Regarding collected rapid responses, the absence of data was itself useful information (eg, the absence of second authors was processed as there being no second authors) and so no analysed data point was considered missing. As there is a lag between submission of a rapid response and publication of the response as a letter, we excluded all rapid responses that were within 66 days of our data collection window (ie, submitted after 16 January 2018). This value was based on preliminary analysis that found a vast majority (80%) of letters were published within 66 days of the rapid response submission date.

Matching protocol

Although both rapid responses and published letters are available freely on BMJ.com, they are available on different parts of the website, and the vast majority of published letters do not link to the specific rapid response that was initially submitted. The task of finding out which rapid responses have been accepted is further complicated by the fact that many editorial changes are made between the submission of a rapid response, and it being printed in the BMJ. Therefore, finding the corresponding rapid response for a letter is not as trivial as looking for a rapid response with identical text content. To carry out this correspondence, a hierarchical matching protocol was used, which we summarise here. For each published letter, we search the corpus of rapid responses for those by the same first author. To make this possible, author names were standardised by removing middle names or initials. When a first author was only associated with a single rapid response, and a single letter to the editor, these were designated as the same submission. When the author of a letter to the editor has submitted numerous rapid responses, one was chosen where the first 50 characters had the highest similarity with the letter to the editor. If no rapid response could be found with the same first author as the published letter to the editor, author name was ignored and rapid responses submitted recently before publication of the letter to the editor were searched for one with the highest similarity in the first 50 characters. A subset of 600 matched rapid responses and letters was checked manually by AB (Alfred Balston) and MZ and found to be 85.3% accurate.

Classifying ethnicity and gender

Authors of rapid responses are not asked to disclose their ethnicity or gender, and the number of rapid responses involved was too great to individually contact the authors and ask this sensitive information. An automated method was used that could determine ethnicity of a name, for many tens of thousands of names, quickly and with little manual input. This took the shape of a previously published machine learning algorithm, nameprism.com, that has been trained to classify ethnicity on 74 million names, as well as being externally validated on datasets other than those which it was trained on.17 To the best of our knowledge, it has demonstrated the highest classification accuracy of any publicly available tool for this task, with an F1 score of 0.795. We follow previous medical research that has used name to classify ethnicity,18 19 as well as a validation study that suggested name analysis is accurate enough to be used to aid health research.20 This ethnicity classification tool was trained on a large, diverse set of names which the authors claim cover 90% of the world’s names.17 It was developed in the USA and so the six ethnicity categories used are American: white, black, Asian and Pacific Islander, Hispanic, American Indian and Alaska Native and more than two races. Ethnicity was estimated using the first and last name of authors, which aims ‘to reduce errors when names are mixtures because of immigration or cross-nationality marriages’. This tool is designed and tuned to infer on a world population, and so is well suited to a journal such as the BMJ with a worldwide authorship. Gender of first author name was determined using a tool called Gender Guesser, which uses a database of approximately 40 000 common names and their corresponding gender.21 The first names of rapid response authors are checked against this database, and placed into one of the following categories: male, female, mostly male, mostly female, androgynous (equal probabilities of being male or female) and unknown (not in the database). In an independent validation, on a manually labelled dataset of 7076 names, it was compared with four other such gender inference tools, and was found to achieve ‘the lowest misclassification rate without parameter tuning for the entire dataset, introducing also the smallest gender bias’.22 We also validated the Gender Guesser tool on a public dataset of 29 872 names extracted from Wikipedia. The tool was able to infer gender for 82.76%. The names inferred as ‘male’ were 99.2% accurate, and those inferred as ‘female’ were 95.6% accurate. Overall, this tool was 98.4% accurate when detecting ‘male’ or ‘female’ names in our validation dataset. The ethnicity and gender classification tools provide an estimated ethnicity and gender that, for the purpose of this study, is assumed to be analogous to the ethnicity and gender that a reader or reviewer would assign to an author.

Statistical analysis

Univariate associations between author and rapid response features, and publication was carried out by calculating χ2 test and t-test scores. Hierarchical binary logistic regression was used to look at the correlation between ethnicity and publication, taking into account other author and rapid response features.

Software used

Gender of first author names was classified using Gender Guesser.21 Ethnicity was classified using nameprism.com.17 Flesch reading ease score was calculated using an open source library called textstat.23 Spelling and grammatical mistakes were quantified using a tool called language-check.24 All code was written in ‘Python 3’ in the ‘Jupyter notebook’ text editor.25 Data collection used an automated script, using the open source Python libraries ‘Requests’ and ‘BeautifulSoup’.26 27 Further processing and data manipulation used the Python libraries ‘NumPy’, ‘Pandas’ and ‘SciKit learn’.28–30 Statistical analysis was carried out in the IBM SPSS V.25 package and in Python.

Results

Baseline data

Analysis was performed on 113 265 rapid responses, of which 8415 (7.4%) were published as letters to the editor. Of all submitted rapid responses, 83% had first authors with names classed as ‘white’; 62% of first authors were classed as ‘male’. See table 1 for baseline author and rapid response features. We also performed an analysis of the characteristics of submissions, broken down by inferred gender and ethnicity. These can be found in online supplemental tables 1 and 2.
Table 1

Characteristics of published and unpublished rapid responses submitted to BMJ.com between 25 April 1998 and 23 March 2018

CharacteristicAll submissions (n=113 265)Published (n=8415)Unpublished (n=104 850)Statistical significance
Author gender (number (%))χ2=181, p<0.0005
 Male70 256 (62.0)5636 (67.0)64 620 (61.6)
 Female18 592 (16.4)1409 (16.7)17 183 (16.4)
 Mostly male2434 (2.1)171 (2.0)2263 (2.2)
 Mostly female1321 (1.2)98 (1.2)1223 (1.2)
 Androgynous1021 (0.9)82 (1.0)939 (0.9)
 Unknown19 641 (17.3)1019 (12.1)18 622 (17.8)
Author ethnicity (number (%))χ2=267, p<0.0005
 White94 077 (83.1)7492 (89.0)86 585 (82.6)
 API15 759 (13.9)726 (8.6)15 033 (14.3)
 Hispanic1903 (1.7)90 (1.1)1813 (1.7)
 Black1204 (1.1)64 (0.8)1140 (1.1)
 AIAN2 (0.0)0 (0.0)2 (0.0)
 Unknown320 (0.3)43 (0.5)277 (0.3)
Word count314±318410±278307±319t=−32.4, p<0.0005
Flesch reading ease50.3±16.347.6±12.150.5±16.5t=20.8 p<0.0005
Has references (number (%))40 173 (35.5)4445 (52.8)35 728 (34.1)χ2=1196, p<0.0005
Number of references1.3±3.02.2±3.31.3±2.9t =−26.6, p<0.0005
Author position (number (%))128, p<0.0005
 Consultant16 291 (14.4)1592 (18.9)14 699 (14.0)
 Professor9959 (8.8)1110 (13.1)8849 (8.4)
 Senior4491 (4.0)523 (6.2)3968 (3.8)
 Student3080 (2.7)143 (1.7)2937 (2.8)
 Other79 444 (70.1)5047 (60.0)74 397 (71.0)
Twitter handle present (number (%))1868 (0.2)254 (0.3)1614 (0.2)χ2=105, p<0.0005
US spelling and grammar errors27.0±44.635.2±39.226.3±44.9t=−19.9, p <0.0005
UK spelling and grammar errors8.9±18.09.4±15.38.9±18.2t=−3.0, p=0.003
Multiple authors (number (%))19 256 (17.0)2914 (34.6)16 342 (15.6)χ2=2002, p<0.0005
Competing interests declared (number (%))6184 (5.5)924 (11.0)5260 (5.0)χ2=537, p<0.0005

Plus–minus values are means±SD. Percentages may not sum to 100% due to rounding.

AIAN, American Indian and Alaska Native; API, Asian and Pacific Islander.

Characteristics of published and unpublished rapid responses submitted to BMJ.com between 25 April 1998 and 23 March 2018 Plus–minus values are means±SD. Percentages may not sum to 100% due to rounding. AIAN, American Indian and Alaska Native; API, Asian and Pacific Islander.

Univariate analysis

Univariate associations were analysed and included in table 1.

Multivariate analysis

All variables above were used in a hierarchical, binary logistic regression with two blocks. The first block included all variables except first author’s gender and ethnicity. The second block additionally contained first author’s gender and ethnicity. First author’s gender and ethnicity remained statistically significant after accounting for measured confounders. In the second block, incorporating this information significantly improved the model (omnibus test of model coefficients, χ2=4648.412, degrees of freedom=43, p<0.0005), the pseudo R2 value was 0.098, up from 0.088 in the first block. Table 2 shows the results of the second, complete logistic regression and ORs for each variable.
Table 2

ORs with 95% CIs and p values in multivariate analysis

VariableBlock 1: ORs (95% CI)Block 1: p valueBlock 2: ORs (95% CI)Block 2: p value
Gender
 Male<0.0005
 Female0.90 (0.85–0.96)0.002
 Mostly male0.97 (0.83–1.14)0.727
 Mostly female0.98 (0.80–1.22)0.882
 Androgynous1.02 (0.80–1.29)0.901
 Unknown0.75 (0.69–0.81)<0.0005
Ethnicity
 White<0.0005
 API0.54 (0.49–0.59)<0.0005
 Hispanic0.51 (0.41–0.64)<0.0005
 Black0.74 (0.57–0.96)0.023
 Unknown1.39 (0.99–1.96)0.057
Word count1.00 (1.00–1.00)<0.00051.00 (1.00–1.00)<0.0005
Flesch reading ease1.00 (1.00–1.00)0.0161.00 (1.00–1.00)0.042
Has references0.71 (0.67–0.75)<0.00050.70 (0.67–0.74)<0.0005
Number of references1.01 (1.00–1.02)0.2901.00 (0.99–1.01)0.475
Author position
 Consultant0.69 (0.65–0.73)<0.00050.70 (0.65–0.74)<0.0005
 Professor0.75 (0.70–0.81)<0.00050.74 (0.69–0.79)<0.0005
 Senior0.73 (0.66–0.81)<0.00050.74 (0.67–0.81)<0.0005
 Student1.64 (1.38–1.95)<0.00051.54 (1.30–1.83)<0.0005
Twitter handle present0.69 (0.60–0.79)<0.00050.69 (0.60–0.79)<0.0005
US spelling and grammar errors1.00 (1.00–1.00)<0.00051.00 (1.00–1.00)0.001
UK spelling and grammar errors1.00 (0.99–1.00)<0.00050.99 (0.99–1.00)<0.0005
Multiple authors0.47 (0.44–0.49)<0.00050.44 (0.42–0.46)<0.0005
Competing interests declared0.56 (0.52–0.60)<0.00050.57 (0.53–0.61)<0.0005
Recent submission1.00 (1.00–1.00)<0.00051.00 (1.00–1.00)<0.0005
Near ideal word count0.54 (0.51–0.57)<0.00050.54 (0.51–0.57)<0.0005
Near ideal Flesch reading ease0.80 (0.76–0.85)<0.00050.80 (0.76–0.85)<0.0005

‘American Indian and Alaska Native’ was removed from the ethnicity figures due to an absence of published letters from that ethnic group.

API, Asian and Pacific Islander.

ORs with 95% CIs and p values in multivariate analysis ‘American Indian and Alaska Native’ was removed from the ethnicity figures due to an absence of published letters from that ethnic group. API, Asian and Pacific Islander.

Discussion

Statement of principal findings

The estimated gender and ethnicity of first author names of BMJ rapid responses were predictive of publication, even when other features of the rapid response and the author were taken into account.

Strengths and limitations of the study

To the best of our knowledge, this is the largest ever analysis of a scientific corpus that looks for associations with publication rate. This was possible because of the open nature of BMJ rapid responses, and through automation at various stages, including data gathering and processing, which includes the use of validated machine learning algorithms for automatic ethnicity classification. This allowed us to analyse over 100 000 submissions, a feat which would not have been possible manually. One of the largest limitations of this study is that it is only sensitive to associations. It is not possible to infer causality from this data, when the exact mechanisms for the observed discrepancy are not known. There could be other unmeasured factors accounting for the discrepancy in publication rates such as subtle differences in communication style, which have been posited to explain at least partly the ethnicity attainment gap in medical school clinical examinations.31 It is worth noting, however, that in clinical examinations it is male students that are consistently found to underperform relative to their female counterparts,32 33 while we found that female first authors were under-represented. Though the tool for classifying ethnicity from name has been validated on a global population, it was developed in the USA and uses ethnicity categories that closely resemble those officially used within the USA. This is not ideal for names outside the USA, where different categories are defined officially. The categorisation of gender resulted in a fairly large proportion (17.3%) of authors with unknown gender, and they were less likely to be published. Another similar tool for inferring gender from name was shown to have an overall 93.8% accuracy in classifying author names in an analysis in the journal Science.34 This high accuracy is due to these techniques’ ability to quantify their uncertainty; for example, if they believe there is a roughly equal chance of the name being male or female, it is classed as ‘androgynous’. Only names which are very likely to be a specific gender are inferred as such. Although there was an option to add location to the gender tool to determine the likely gender of a name in a specific country, this was not done for two reasons. First, the location data extracted from rapid responses was highly heterogeneous with some authors providing countries, cities or institution names, or multiple addresses, without consistent spelling or abbreviations. Second, of the 45 376 names in the Gender Guesser dictionary, only 286 (0.6%) names are influenced sufficiently by location such that the estimate is changed from ‘male’ to ‘female’ or vice versa. In this study, gender and ethnicity were estimated from the author’s name, which provides a proxy for the gender and ethnicity that a reviewer would assign to an author given the same information. While assigning ethnicity and gender based on name may not always match the self-identified ethnicity and gender of the author, it is practical and necessary in this setting. A difficulty in classifying both ethnicity and gender into discreet categories is that the nuances of these complex social identities is lost, for example, in the gender tool there is no categorisation for non-binary genders or transgender individuals, and limited provision for mixed race individuals in the ethnicity tool. A limitation in this work is that for all analysed letters, the corresponding rapid response had to be imputed using the protocol mentioned previously. A small minority of letters seem not to have been submitted as rapid responses, which may represent either direct publications from the editor or direct correspondences between a paper author and the editor. It is worth noting, however, that recently published letters link directly to the original rapid response that was submitted. This would allow future analysis to have ground truth data on which rapid responses were published and which were not.

Strengths and limitations in relation to other studies

A recent report by Science found no statistically significant evidence that their editorial process under-represented female authors.35 In this report, gender was identified manually, which limited their sample size to a small random selection of submissions. The gender disparity we observed is relatively subtle compared with that which we observed for ethnicity, and this might mean that larger sample sizes are needed to elucidate any discrepancies. The under-representation of ethnicity has so far been less thoroughly studied in comparison to gender. Nonetheless, our findings are in line with published data that BME authors are under-represented in published articles10; however, our study is further able to identify that submissions from BME authors are less likely to be accepted for publication relative to similar submissions by their white counterparts.

Implications

In our hierarchical logistic regression, gender and ethnicity explain a small amount of additional variance (0.01 increase in the pseudo R2) compared with the other features alone; however, this is considerable compared with a low pseudo R2 baseline of 0.088. Given the likely complexity of the selection process, it is unsurprising that the pseudo R2 is low, as there are many unmeasured factors. Indeed, publication is determined by the expert opinion of the editor, on things that are impractical or impossible to quantify in a study like this, including clarity, style and interest to the potential reader. Nonetheless, one factor which may play a role is that of unconscious gender or racial bias. Implicit bias has been documented in clinical decision-making,36 medical school admissions37 and selection of junior doctors.38 It is important to state, however, that the current study design does not provide causal evidence of bias, which would require a prospective experimental study design to fully account for other unmeasured factors. Our results suggest that scientific journals should look for such discrepancies in all forms of submissions, including opinion pieces and research papers, which are not posted publicly. Such analysis should include looking at differences in submission and publication rates by author ethnicity as well as by author gender. This study demonstrates the ability of data science and machine learning techniques to rapidly extract and analyse a large and complex dataset with relative ease. Without these techniques, this analysis would not have been possible. Being able to automate the process of feature extraction, including gender and ethnicity, opens avenues for further observational studies of open access data. It has also opened countless possibilities in medicine across the entire patient journey, from triage and improving attendance, to automated disease diagnosis, prognostication, management and even the discovery or repurposing of new medications.39–45

Unanswered questions and future research

This work highlights important associations in past data; however, more research is necessary to draw concrete conclusions regarding the reasons for these associations. For example, other confounders might be considered, including communication style, field of study of the rapid response author and of the article being responded to, and the locations of institutes of submitted pieces. It may also be interesting to see how these discrepancies change over time, if at all. Additionally, studies have demonstrated unconscious gender biases in science,12 and unconscious racial biases in other areas,36–38 but far less research has studied unconscious racial biases in academia. It is important to establish whether the discrepancies we found in BMJ letters to the editor are present in other journals, and for other scientific manuscript types such as original research. Though trends have been studied for published papers, quantifying the rate of acceptance is an invaluable way to eliminate the confounder that is the number of submissions, and we hope that future research in this field can either be done by journals themselves, or by researchers in close collaboration with journals to ensure that this submission data is included in any analysis.

Conclusion

A number of variables were identified that correlated with the acceptance rate of rapid responses. Discrepancies in the publication rates between genders and ethnicities remained significant after accounting for other factors. The cause of these discrepancies is unclear and may in part be explicable by implicit bias. Regardless of the cause, it is evident that female and BME voices are under-represented, and efforts should be made to identify these causes and rectify them.
  23 in total

Review 1.  Big data and machine learning algorithms for health-care delivery.

Authors:  Kee Yuan Ngiam; Ing Wei Khor
Journal:  Lancet Oncol       Date:  2019-05       Impact factor: 41.316

2.  Looking inward at gender issues.

Authors:  Jeremy Berg
Journal:  Science       Date:  2017-01-27       Impact factor: 47.728

3.  Racial discrimination against doctors from ethnic minorities.

Authors:  A Esmail; S Everington
Journal:  BMJ       Date:  1993-03-13

4.  Women in Academic Medicine Leadership: Has Anything Changed in 25 Years?

Authors:  Paula A Rochon; Frank Davidoff; Wendy Levinson
Journal:  Acad Med       Date:  2016-08       Impact factor: 6.893

5.  A matter of priorities? Exploring the persistent gender pay gap in hospital medicine.

Authors:  A Charlotta Weaver; Tosha B Wetterneck; Chad T Whelan; Keiki Hinami
Journal:  J Hosp Med       Date:  2015-06-30       Impact factor: 2.960

6.  Name analysis to classify populations by ethnicity in public health: validation of Onomap in Scotland.

Authors:  F Lakha; D R Gorman; P Mateos
Journal:  Public Health       Date:  2011-09-09       Impact factor: 2.427

7.  Characterizing Race/Ethnicity and Genetic Ancestry for 100,000 Subjects in the Genetic Epidemiology Research on Adult Health and Aging (GERA) Cohort.

Authors:  Yambazi Banda; Mark N Kvale; Thomas J Hoffmann; Stephanie E Hesselson; Dilrini Ranatunga; Hua Tang; Chiara Sabatti; Lisa A Croen; Brad P Dispensa; Mary Henderson; Carlos Iribarren; Eric Jorgenson; Lawrence H Kushi; Dana Ludwig; Diane Olberg; Charles P Quesenberry; Sarah Rowell; Marianne Sadler; Lori C Sakoda; Stanley Sciortino; Ling Shen; David Smethurst; Carol P Somkin; Stephen K Van Den Eeden; Lawrence Walter; Rachel A Whitmer; Pui-Yan Kwok; Catherine Schaefer; Neil Risch
Journal:  Genetics       Date:  2015-06-19       Impact factor: 4.562

8.  Implicit Racial Bias in Medical School Admissions.

Authors:  Quinn Capers; Daniel Clinchot; Leon McDougle; Anthony G Greenwald
Journal:  Acad Med       Date:  2017-03       Impact factor: 6.893

9.  Predicting scheduled hospital attendance with artificial intelligence.

Authors:  Amy Nelson; Daniel Herron; Geraint Rees; Parashkev Nachev
Journal:  NPJ Digit Med       Date:  2019-04-12

10.  A Deep Learning Approach to Antibiotic Discovery.

Authors:  Jonathan M Stokes; Kevin Yang; Kyle Swanson; Wengong Jin; Andres Cubillos-Ruiz; Nina M Donghia; Craig R MacNair; Shawn French; Lindsey A Carfrae; Zohar Bloom-Ackermann; Victoria M Tran; Anush Chiappino-Pepe; Ahmed H Badran; Ian W Andrews; Emma J Chory; George M Church; Eric D Brown; Tommi S Jaakkola; Regina Barzilay; James J Collins
Journal:  Cell       Date:  2020-02-20       Impact factor: 41.582

View more
  1 in total

1.  The Under-representation and Stagnation of Female, Black, and Hispanic Authorship in the Journal of the American Medical Association and the New England Journal of Medicine.

Authors:  Moustafa Abdalla; Mohamed Abdalla; Salwa Abdalla; Mohamed Saad; David S Jones; Scott H Podolsky
Journal:  J Racial Ethn Health Disparities       Date:  2022-03-21
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.