Literature DB >> 35496681

Participation and Performance by Gender in Synchronous Online Lectures: Three Unique Case Studies during Emergency Remote Teaching.

Sierra C Nichols1, Yongyong Y Xia1, Mikaylie Parco1, Elizabeth G Bailey1.   

Abstract

Studies have documented that men's voices are generally heard more than women's voices in face-to-face undergraduate biology classes, and some performance gaps have also been documented. Some of the few studies on gender equity in traditional online biology education suggest that women participate more and perform better in asynchronous online courses compared to men. While much is known about emergency remote teaching during the COVID-19 pandemic exacerbating existing inequities generally, studies are needed about the impact of specific emergency remote teaching practices on specific groups such as women. In this study, we performed an in-depth investigation of three life sciences classrooms that utilized synchronous online lectures during the pandemic. We observed each class throughout the semester, quantified participation behaviors, and investigated the role of student gender. We also compared final course grades by gender. On average, we found that men participated more than women both verbally and by chat. These differences were not significant for each class individually, but the differences align with the face-to-face patterns seen in this population previously. Our results also hint that men's chat comments may be more likely to be acknowledged than women's chats by peers. We found evidence of greater performance disparities favoring men than seen previously before the pandemic, but not in all classes. We discuss implications for instructors conducting emergency remote teaching as well as the need for larger studies to test the replicability of our results.
Copyright © 2022 Nichols et al.

Entities:  

Keywords:  COVID; Zoom; achievement; biology education; female; gender gaps; undergraduate; verbal participation; videoconference; women in STEM

Year:  2022        PMID: 35496681      PMCID: PMC9053025          DOI: 10.1128/jmbe.00281-21

Source DB:  PubMed          Journal:  J Microbiol Biol Educ        ISSN: 1935-7877


INTRODUCTION

Women are underrepresented in science, technology, engineering, and math (STEM), but these fields have become more equitable over time. In the last 50 years, the number of women in the workforce in STEM fields increased from 8% to 27% (1). Women are well represented in biology, receiving more than half of the degrees awarded in 2016 at the bachelor’s, master’s, and doctorate levels (2). However, a closer look at the natural sciences reveals subtle differences referred to as the leaky pipeline, as we see fewer women represented along the path toward tenured professorship (3–6). At the same time, demand for educated and diverse workers is continually increasing, and increased inclusion of women could help meet that demand (7). Research suggests that gender diversity benefits companies and organizations, because women’s unique experiences are likely to give them different perspectives and ideas (8, 9). Thus, it is important that we strive to make STEM education more equitable and close opportunity gaps for women (7). As discussed by Eddy and Brownell (10), equal enrollment numbers do not guarantee gender parity, and the effects of undergraduate student performance, engagement, and student affect have important implications down the pipeline. For example, women may recognize gender inequities in classroom participation more than men and thus note classroom environments that are more unwelcoming (11). These classroom experiences can have long-lasting effects, as women are less visible and lack social standing when they do not speak up (12) and students’ sense of belonging in STEM impacts their STEM career interests (13). Furthermore, participation can impact performance, as students’ perceptions of engagement, including active or collaborative learning and faculty interaction, are strongly related to their grades (14). Again, classroom performance can have long-term effects as grades and perceptions of ability or performance are strong predictors of persistence (15, 16). Thus, we chose to focus on undergraduate women’s in-class participation and course performance.

Participation and performance gender inequities in undergraduate biology classrooms

Recent studies on traditional in-person biology classrooms have documented that men are more likely to speak up than women in class (17–21). Some studies have found performance gender gaps in favor of men in introductory biology classes, especially on exams requiring higher-level cognitive skills (18, 22), while others found either no difference by gender (23) or mixed results (24). We previously found that gender gaps in performance varied across a wide range of life sciences courses, and that women were more likely to underperform compared to men in courses instructed by a man and/or where women students were the minority (17). A recent meta-analysis found that large class sizes, high reliance on exam scores, and lack of active learning strategies were all predictive of decreased women’s performance compared to men in undergraduate biology and chemistry classes (25).

Emergency remote teaching during the covid-19 pandemic

During the global COVID-19 pandemic, universities canceled face-to-face classes and transitioned to online learning. This instruction has been termed emergency remote teaching (ERT) to contrast this sudden and temporary change from traditional online learning (26). Students have reported some benefits associated with ERT such as flexibility (27, 28), but the literature also lists many challenges (27–31). We note that difficulties arising from the pandemic were not felt by all populations equally, and existing inequalities were exacerbated (31–34). Thus, we must investigate how ERT affected marginalized underrepresented groups in STEM, including women.

Gender inequities in ERT and traditional online learning

Recent research found that women were more likely to report decreased motivation, focus, and self-direction than their male peers during ERT (31). Furthermore, studies suggest that women are more likely to turn off their camera due to concerns about their background (35), and women may get zoom fatigue due to mirror anxiety (i.e., anxiety triggered by the perpetual mirror during video conferences) more than men (36). Reinholz et al. (37) studied biology classrooms that started in person and then transitioned online mid-semester and did not find evidence of widened gender inequities. They suggest that the diverse Zoom features provided more participation spaces for women. Much of the previous research on traditional online learning has been done on courses that utilize asynchronous activity, and most results suggest that women have a more positive perception of online classes than males, especially in terms of communication and connectedness, and women often perform better than men online (38–40). In a large study that compared students’ online performance to previous in-person experiences, researchers found that students generally performed worse in online settings, especially men (41). Many women described anonymity and asynchronous aspects of the course as the most positive aspects (42). Many studies on asynchronous online discussions report that women post more than men and show higher engagement and collaboration (43, 44); however, some studies have found no gender differences in post rates (45).

Our research questions

While there is a growing body of research on gender inequities in performance and in-class participation for biology classes, the impact of ERT practices on gender equities in undergraduate biology classes is largely unknown. We chose to investigate synchronous online lectures, as these are the most like face-to-face classroom experiences, and performed an in-depth analysis of three life sciences classrooms to investigate these research questions: How are instructors inviting participation in synchronous online lectures during ERT? Are men and women equally likely to participate in synchronous online lectures overall? Verbally? By chat? Does this depend on the gender ratios in the class? Do these participation trends change over the course of the semester? Are there gender differences in final course grades earned in ERT contexts that use synchronous lectures?

METHODS

Ethics statement and participant inclusion criteria

This study was approved by the Brigham Young University Institutional Review Board (Protocol E2020-356). After receiving approval, we emailed the instructors of 61 synchronous remote classes in the life sciences during the 2020–2021 school year when ERT was common. Fourteen of those instructors agreed to let us seek consent from their students to consider their class, and all students consented for Zoom recordings to be released in six of those classes. Three of those six classes were excluded because there were fewer than three men enrolled. The final inclusion criteria were that the course must be an undergraduate class in the life sciences, utilize live/synchronous online lectures, have at least three men enrolled, and the instructor and all students needed to consent for videos to be viewed. While we originally intended to replicate the scale of our previous study on in-person classrooms (17) (n = 34), we ended up with a smaller number of classes (n = 3). This allowed us to do a much more in-depth study of these classes.

Study population

Our study took place at a private, religious R2 university. The student population is broadly Christian, relatively conservative, and generally high performing due to the competitive nature of enrollment. In the college, ∼50% of the students are women and ∼15% of the faculty are women (internal data, 2019). The course characteristics for the three classes used in this study are summarized in Table 1, and more details about each course are found in the Supplemental Materials. In Table 1, “points awarded subjectively” refers to points earned on open response questions, essays, etc. as opposed to multiple choice or similar question types that are computer graded. These subjective points could be on exams or lower-stakes homework assignments.
TABLE 1

Course characteristics

CharacteristicClass 1Class 2Class 3
Class level400-level400-level100-level
AudienceMajorsMajorsGeneral Education
Instructor genderFemaleMaleMale
Teaching assistent gendersFemaleFemaleFemale
Class size47 students22 students67 students
 Female25 (53%)13 (59%)48 (72%)
 Male22 (47%)9 (41%)19 (28%)
Chat/verbal participation required?NoYesNo
Percent of final grade determined subjectively∼33%∼66%∼70%
Percent of final grade dependent on exam scores50%43%30%
Course characteristics

Classroom observation methods

We observed class sessions by analyzing video recordings and chat box text files from Zoom videoconferencing software. We collected recordings of three lectures that took place early in the semester (first 5 weeks of a 15-week semester), three from mid-semester, and three from late in the semester (last 3 weeks of semester). For Class 1, we were unable to get recordings early in the semester, so only mid-semester and late semester data are included. For some analyses, results from all six (Class 1) or nine (Class 2 and Class 3) observations were averaged or combined to get an overall look at the whole semester. In other analyses, results from three observations were averaged or combined to give an average look at participation at a specific point in the semester (early, mid, or late).

COPUS

First, we used the Classroom Observations Protocol for Undergraduate STEM (COPUS) to analyze each recording (46, 47). Twenty-five unique codes were used to quantify instructor and student behaviors in 2-min intervals. We also collapsed those 25 codes into four codes for instructors and four codes for students (48) (see Appendix 1). Two researchers (S.C.N. and Y.Y.X.) watched and coded the first video for each class individually, and we assessed interrater reliability. Some of the COPUS codes were never coded by our researchers (e.g., students presenting to the entire class). Cohen’s kappa for the rest of the 25 codes ranged from 0.835 to 1.000. Cohen’s kappa for the eight collapsed codes ranged from 0.913 to 1.000. As raters had excellent agreement, each subsequent video was then only coded by one researcher. We report COPUS results in a primarily descriptive manner with collapsed codes (48), as we did not have a large data set on which we could perform cluster analysis as has been suggested (46). We also looked at specific codes, including percent of time spent lecturing, to compare our data to previous cluster analyses which grouped courses into didactic, interactive lecture, and student-centered classrooms (49). However, we recognize the limitations of using the cluster analysis from a separate data set and merely use these categories to discuss the differences between our three classes.

Instructor invitations and student participation

Next, we quantified student participation and instructor invitations to participate based on a protocol adapted from our previous in-person classroom observations (17). For each video, we recorded the number of times the instructor made an explicit invitation for verbal participation, invited students to use the chat feature, put students in breakout rooms, and polled the class. More detailed definitions of these behaviors are found in Appendix 1. We also recorded the number of times each student typed in the chat box or participated verbally. We also noted whether students’ chat comments were acknowledged (either verbally or in another chat comment) by instructors (the course instructor or TAs) or peers. In the recordings, we could not see a participant list, so we could not tell whether students on the roster were logged into Zoom or not. Thus, we did not have a clear way to track absence vs silence, so non-participators in our study may either be absent students or attending students who never participated. More detailed definitions of these behaviors are found in Appendix 1. We assumed and recorded each student’s gender based solely on whether they were male- or female- presenting in photos or their voice. This is clearly a limitation of our study, as we do not know the possible complexities of each student’s gender identity. The numbers of men and women in each class determined in this way matched the numbers of males and females in the lists obtained from the registrar’s office, which include the sex listed on students’ educational records (see Table 1). When recording settings did not save student names, we matched verbal participators to the pictures provided on the course picture roll. If the face was not an obvious match, we consulted with the research team to come to a consensus on which student was participating. Because we had six or nine observations of each class, students who participated with their cameras off occasionally were able to be identified based on voice comparisons with other observations in which their camera was turned on. Three raters (Y.Y.X., S.C.N., M.P.) analyzed the first video for each class. Initially, interrater reliability (Cohen’s kappa) ranged from 0.347 to 0.762. Raters then met and talked through differences until they came to a consensus. The three raters then all rated the same video and achieved near perfect alignment (Y.Y.X. and S.C.N. kappa = 0.98; Y.Y.X. and MP kappa = 0.96; S.C.N. and M.P. kappa = 0.98). Most of the rest of the videos were then coded by a single rater. Occasionally a video had so much participation that two raters would work together to properly identify all students.

Course grades

We obtained final letter grades for all students in each of the three courses from the university registrar’s office. Data were de-identified except for gender (biological sex on educational records) and ACT score. Thus, we can compare final course performance by gender overall for each course, but we cannot match specific students’ participation behavior with their course performance. Final letter grades (e.g., A, A–, B+, etc.) were converted to a 4-point scale (e.g., A = 4, A– = 3.7, B+ = 3.3, etc.). When grades are presented as ACT-adjusted grades, these are the estimated marginal means for gender after performing a one-way ANCOVA. The one-way ANCOVA was performed for each class separately, with gender as the independent variables and ACT score as the covariate as an imperfect measure of preparation.

RESULTS

Course characteristics

First, we analyzed video recordings of synchronous zoom lectures using COPUS (see Methods). As shown in Fig. 1A and B, the three courses we observed were all unique in terms of teaching style. Class 1 was more didactic with ∼70% of instructor codes fitting in the “presenting” category (Fig. 1A; blue). Class 2 was more student-centered with ∼70% of instructor codes fitting in the “guiding” category (Fig. 1A; orange). Finally, Class 3 was more of an interactive lecture, but there was more variety from class period to class period. On average, student codes were evenly split between “receiving” and “sharing in front of class” (Fig. 1B; blue and yellow, respectively).
FIG 1

Teaching style based on instructor and student behaviors. Panels A and B: Zoom recordings of synchronous lectures were coded using the Classroom Observation Protocol for Undergraduate STEM (COPUS). All codes were collapsed into four categories for instructor behaviors (Panel A) and four categories for student behaviors (Panel B). The x axes represent the percent of all codes. Data are averaged from six (Class 1) or nine (Classes 2 and 3) observations, and error bars represent standard error of the mean. Panel C: Instructor behaviors that invite student participation were quantified for each video recording (n = 6 for Class 1 and n = 9 for Class 2 and Class 3).

Teaching style based on instructor and student behaviors. Panels A and B: Zoom recordings of synchronous lectures were coded using the Classroom Observation Protocol for Undergraduate STEM (COPUS). All codes were collapsed into four categories for instructor behaviors (Panel A) and four categories for student behaviors (Panel B). The x axes represent the percent of all codes. Data are averaged from six (Class 1) or nine (Classes 2 and 3) observations, and error bars represent standard error of the mean. Panel C: Instructor behaviors that invite student participation were quantified for each video recording (n = 6 for Class 1 and n = 9 for Class 2 and Class 3). We quantified instructor behaviors that encouraged participation for each observation. As shown in Fig. 1C, instructors of all three courses used a lot of invitations for students to unmute and participate verbally (green), but they rarely explicitly directed the students to use the chat feature. Only the instructors of Class 2 and Class 3 utilized Zoom breakout rooms (red = # sessions; orange = % class time spent in breakout rooms). None of the instructors used Zoom polls frequently (blue), and we noted that the instructors of Class 2 and Class 3 only used these in the observations that took place early in the semester.

Student performance

We acquired final course grades for each class from the registrar’s office to compare student performance by gender. In terms of raw grades, male students significantly outperformed their female peers in Class 1 (Fig. 2A P = 0.04 by unequal variance t test), and this difference of about a half of a letter grade was virtually unchanged after adjusting for ACT score (Fig. 2B P = 0.05 by unpaired t test). In Class 2, males also outperformed females in the course by about a full letter grade (Fig. 2C), but the difference was only significant for ACT-adjusted scores (Fig. 2D P = 0.007 by unpaired t test). Finally, males and females earned equivalent grades in Class 3 (Fig. 2E and F).
FIG 2

Final course grades by gender. Final course grades for Class 1 (Panels A and B), Class 2 (Panels C and D), and Class 3 (Panels E and F) were obtained from the university registrar and were converted to a four-point scale. Panels A, C, and E show raw grade distributions (minimum to maximum with quartiles) by gender. Panels B, D and F show estimated marginal means after adjusting for ACT score (mean ± standard error). Brackets show P value after unpaired t test for each panel, with Panels A and E requiring an unequal variances t test.

Final course grades by gender. Final course grades for Class 1 (Panels A and B), Class 2 (Panels C and D), and Class 3 (Panels E and F) were obtained from the university registrar and were converted to a four-point scale. Panels A, C, and E show raw grade distributions (minimum to maximum with quartiles) by gender. Panels B, D and F show estimated marginal means after adjusting for ACT score (mean ± standard error). Brackets show P value after unpaired t test for each panel, with Panels A and E requiring an unequal variances t test.

Student participation

Next, we quantified student participation from recordings of synchronous Zoom lectures. As shown in Fig. 3, overall participation and gender comparisons varied by class. In Class 1, about 60% of the students never participated, but this was true for both men and women (Fig. 3A). Those that did participate were equally likely to choose chat or verbal participation regardless of gender (Fig. 3A). Furthermore, we saw no difference in the average number of chat comments (Fig. 3B) or verbal participation events (Fig. 3C) by gender in Class 1, although the students who participated the most were men (Fig. 3B and C). All participation in Class 2 was verbal for almost all men and women, and the only two students who never participated were women (Fig. 3D). Both chat and verbal event rates were indistinguishable by gender (Fig. 3E and F), but chatting was very rare in this class. Class 3 utilized both the chat feature and verbal participation (Fig. 3G), and women may have been more likely to choose chat over verbal participation and more likely to not participate at all, χ2(3) = 5.12 (P = 0.16). However, based on average number of events per observation, men and women were equally likely to chat (Fig. 3H) and participate verbally (Fig. 3I). Interestingly, the students who participated the most in this class were women (Fig. 3I). For comparison to other studies (17, 20), the rate ratios (average women’s rate/average men's rate) are 0.30, 0.70, 0.12, 0.63, 0.78, and 0.91 (Fig. 3B, C, E, F, H, and I, respectively).
FIG 3

Classroom participation by gender. All instances of student participation (typed chats or verbal) were recorded from videos of synchronous Zoom lectures for Class 1 (Panels A–C), Class 2 (Panels D–F), and Class 3 (Panels G–I). Panels A, D, and G: Students were classified as participating both verbally and by chat at least once each over all observations (“Both”), at least once verbally but never by chat (“Verbal”), at least once by chat but never verbally (“Chat”), or never participating in any observation (“None”). Fraction of students in each category is shown on the y axis by gender. Panels B, E, and H: For each student, chat rate was calculated as the average number of chat comments per observation. Panels C, F, and I: For each student, verbal event rate was calculated as the average number of verbal comments per observation. Rates are compared by gender, and brackets show P value after unequal variances t test (Panels B, C, E, and I) or unpaired t test (Panels H and F).

Classroom participation by gender. All instances of student participation (typed chats or verbal) were recorded from videos of synchronous Zoom lectures for Class 1 (Panels A–C), Class 2 (Panels D–F), and Class 3 (Panels G–I). Panels A, D, and G: Students were classified as participating both verbally and by chat at least once each over all observations (“Both”), at least once verbally but never by chat (“Verbal”), at least once by chat but never verbally (“Chat”), or never participating in any observation (“None”). Fraction of students in each category is shown on the y axis by gender. Panels B, E, and H: For each student, chat rate was calculated as the average number of chat comments per observation. Panels C, F, and I: For each student, verbal event rate was calculated as the average number of verbal comments per observation. Rates are compared by gender, and brackets show P value after unequal variances t test (Panels B, C, E, and I) or unpaired t test (Panels H and F). We were also able to track the participation of each student over time, and these detailed data are shown in Appendix 2 (Fig. S1–S3). These heat maps show the complex landscape of class participation as a function of time. Class 1 participation is dominated by a few students, especially one man (Fig. S1). Class 2 has much more spread-out participation among most students, although one man and one woman still dominate the discussion (Fig. S2). Class 3 is somewhat in the middle with many silent students (mostly women), and a few women who were the highest participators (Fig. S3). We noticed that in Fig. S1, only two women participated in any of the three mid-semester observations, but many women participated later in the semester. We quantified the proportion of men and women who participated at least once at different phases in the semester, and the data are shown in Fig. 4. By Fisher’s exact test, women were significantly less likely to participate than men mid-semester (P = 0.03) but perhaps more likely to participate than men later (P = 0.20). Our results also would have looked different had we only looked at a specific point in the semester for the other classes. In Class 2, women participated less and less as the semester went on (Fig. S2), but this could also be true for men. In Class 3, women may have been least represented in classroom participation in the middle of the semester (Fig. S3).
FIG 4

Class 1 participation representation by gender changed during semester. The fraction of students who participated at least once (including chats and verbal participation) was calculated based on Zoom lecture video observations. Mid-Semester data came from the three observations in the middle of the semester, and Late in Semester data came from the three observations near the end of the semester. Students had to participate at least once in any of the three lectures to be counted in the ≥ 1x category. Men were more likely to participate mid-semester (Fisher’s Exact Test: P = 0.03, n = 45), but women participated more later in the semester (Fisher’s Exact Test: P = 0.20, n = 45).

Class 1 participation representation by gender changed during semester. The fraction of students who participated at least once (including chats and verbal participation) was calculated based on Zoom lecture video observations. Mid-Semester data came from the three observations in the middle of the semester, and Late in Semester data came from the three observations near the end of the semester. Students had to participate at least once in any of the three lectures to be counted in the ≥ 1x category. Men were more likely to participate mid-semester (Fisher’s Exact Test: P = 0.03, n = 45), but women participated more later in the semester (Fisher’s Exact Test: P = 0.20, n = 45). Finally, we investigated whether men’s and women’s chat comments were equally likely to be acknowledged by instructors and peers. Due to the limited use of the chat feature in Classes 1 and 2 (see Fig. 3), we focused on Class 3. As shown in Fig. 5A, chats of men and women were equally likely to be acknowledged by instructors. However, we saw a hint that men’s chats may get acknowledged more frequently by peers than women’s chat comments (Fig. 5B), with a few men’s chats getting acknowledged more frequently than anyone else’s. We saw no association between acknowledgments and future participation rates (data not shown).
FIG 5

Average number of times students’ chats were acknowledged by instructors or peers in Class 3. For all students who chatted at least once during the nine observations of Class 3 (31 women and 15 men), we calculated the average number of times their chat was acknowledged by instructors (Panel A; includes both the primary instructor and the TAs) or peers (Panel B). Distributions (minimum to maximum with quartiles) are shown by gender. Brackets show P value after unpaired t test (Panel A) or unequal variances t test (Panel B).

Average number of times students’ chats were acknowledged by instructors or peers in Class 3. For all students who chatted at least once during the nine observations of Class 3 (31 women and 15 men), we calculated the average number of times their chat was acknowledged by instructors (Panel A; includes both the primary instructor and the TAs) or peers (Panel B). Distributions (minimum to maximum with quartiles) are shown by gender. Brackets show P value after unpaired t test (Panel A) or unequal variances t test (Panel B).

DISCUSSION

We did an in-depth analysis of ERT in three life sciences classrooms that had synchronous lectures. These three courses were very different in context, composition, and instructional style (Table 1, Fig. 1). We can compare results from this case study of ERT to those from in-person courses published on our population previously (17). With women making up about 70% of students enrolled in male-instructed Class 3, the lack of a gender difference in final course grades (Fig. 2E and F) aligns with what we would predict for in-person courses in our population (see Fig. 6 in Ref. 17). However, we saw much larger gender gaps in final course grades for Class 1 (about a half letter grade in favor of males; Fig. 2A and B) and Class 2 (about a full letter grade in favor of males, Fig. 2C and D). From our previous data set (see Fig. 6 in Ref. 17), we would have predicted females to earn higher grades than males in a female-instructed class with about 50% female students like Class 1, and we would have predicted no gender gap in a male-instructed class with about 60% female students like Class 2. This suggests that ERT classrooms may exacerbate gender achievement gaps. Class 1 results would align with our previous results if it had been taught by a man, so perhaps the benefits of female role models are reduced online without in-person relationships. Our results do align with a recent meta-analysis (25) in that Classes 1 and 2 had greater dependence on exam scores (Table 1) and underperformance of women (Fig. 2A–D). Our results differ from some studies in traditional online learning that found women perform better than men (40). Unsurprisingly, more active classrooms (Classes 2 and 3; Fig. 1) provided opportunities for a larger proportion of all students to be involved and participate in class (Fig. 3D, G, Fig. S2 and S3). Our most didactic class (Class 1; Fig. 1) had a much smaller proportion of students participating (Fig. 3A and Fig. S1). We saw no clear association between how active a classroom was and gender equity in participation, as women may have been more likely to stay silent than men in Class 3 (Fig. 3G), but no clear gender pattern was seen in Class 1 (Fig. 3A). This could possibly be explained by the slightly larger class size in Class 3 (Table 1); (20), but Class 3 also had more women, which has been shown to predict more gender equity (17). Perhaps having more female peers adds belonging in person when these peers are consistently visible, but invisible female peers in an online lecture may not have the same effect. The degree to which female peers are “invisible” would depend on camera use in the class and whether students were visible to each other. Unfortunately, we could not quantify the fraction of cameras that were on because Zoom recordings did not show the full gallery view when the instructor was sharing their screen (which applied most of the time). In terms of in-class participation rates, no class showed statistical differences between men and women (Fig. 3); however, in our previous study on this population, small participation gaps were statistically reproducible over many classes and predicted by student gender ratios (17). If we compare our rate ratios (women’s average rate/men’s average rate) with those in that original study (see Fig. 4A in Ref. 17) for classes with 50% females (Class 1), 60% females (Class 2), and 70% females (Class 3), our verbal rate ratios of 0.7, 0.63, and 0.91, respectively, align well with those original results. Thus, we did not find any evidence that gender gaps in verbal participation are any different for synchronous online lectures versus in-person lectures. Interestingly, the chat rate ratios (women/men) were much lower than our in-person verbal data from 2020 for Class 1 (0.30) and Class 2 (0.12), suggesting there could be larger gender gaps in favor of men for typed chats. However, using the chat feature was not the norm in these classes, and chat events were very rare. In Class 3, where chatting was the norm, the chat rate ratio (women/men) of 0.78 aligns well with our previous data. Thus, we believe that the chat feature is as equitable as verbal participation if it is commonly used according to class norms. If we look at the fraction of students who used the chat and/or unmuted in Class 3 (where both methods were common; Fig. 3G), we see hints that women may choose the chat as opposed to participating verbally. As proposed previously (37), using a variety of different Zoom features could possibly facilitate a more friendly and gender equitable environment. Women have been shown to use asynchronous communication channels more than men in traditional online learning to create online community (43, 44), so it is possible women are doing the same thing with the chat feature. In our previous study (17), participation and performance gender gaps were not associated with each other. Here, we saw a possible correlation between participation gender equity (Fig. 3) and performance gender equity (Fig. 2), as Class 3 had the least difference between grades by gender and the participation rate ratio closest to 1. As discussed above, there are other factors that we would expect to affect grade differences by gender, but participation trends in class might also contribute. Our study design allowed us to look systematically at different points in the semester and observe each class six to nine times. Previous studies have observed classes anywhere from two to 20 times, but most classes were observed closer to three times each (17–20). We found that conclusions on gender equity in participation may differ based on how many observations you make and at what point in semester you make them, as we would have drawn different conclusions with fewer observations at specific points in semester (Fig. 4, Fig. S1–S3). Finally, we saw hints that men’s chat comments may be acknowledged by peers more than those made by women (Fig. 5B). This aligns with research that suggests men are likely to overestimate academic ability of male peers and underestimate ability of female peers (50). Male and female faculty can also have biases that favor men (51). However, we have not seen evidence of faculty explicitly favoring men's participation in the past (17), nor did we see instructors acknowledging men’s or women’s chats differently here (Fig. 5A).

Limitations

First, we acknowledge that we were unable to have students self-disclose their gender identity. Because most students did not have their pronouns labeled on their display name, we had to rely on stereotypical features to classify students as male- or female-presenting. Thus, we do not know the complexities with which students identify based on gender. Furthermore, we obtained de-identified grades from the registrar’s office which were categorized according to biological sex, which we acknowledge may or may not match the gender which students present. Our conclusions are limited by our inability to differentiate between absent students and attending students who remain silent. As ERT may lead to less consistent class attendance, gender differences in attendance habits could confound our results. Our small sample size is also a limitation. Future replication studies should help identify which of our results are generalizable patterns.

Implications for instructors

Based on our results, we encourage instructors to have more interactive and student-centered classrooms. Utilizing a greater variety of Zoom features (e.g., breakout rooms, polls, and the chat feature) will likely increase the proportion of students who get to participate, and women may be more likely to participate via chat, especially in large classrooms. Anecdotally, we noted that longer pauses may be necessary in a Zoom lecture compared to in-person lectures, as students may take extra time to both get up the courage to speak and to unmute themselves. We also saw the participation trends did change over time, for better or worse, so it is never too late in the semester for instructors to re-evaluate their pedagogical choices and encourage broad and diverse participation.
  16 in total

1.  Anatomy of STEM teaching in North American universities.

Authors:  M Stains; J Harshman; M K Barker; S V Chasteen; R Cole; S E DeChenne-Peters; M K Eagan; J M Esson; J K Knight; F A Laski; M Levis-Fitzgerald; C J Lee; S M Lo; L M McDonnell; T A McKay; N Michelotti; A Musgrove; M S Palmer; K M Plank; T M Rodela; E R Sanders; N G Schimpf; P M Schulte; M K Smith; M Stetzer; B Van Valkenburgh; E Vinson; L K Weir; P J Wendel; L B Wheeler; A M Young
Journal:  Science       Date:  2018-03-29       Impact factor: 47.728

2.  Science faculty's subtle gender biases favor male students.

Authors:  Corinne A Moss-Racusin; John F Dovidio; Victoria L Brescoll; Mark J Graham; Jo Handelsman
Journal:  Proc Natl Acad Sci U S A       Date:  2012-09-17       Impact factor: 11.205

3.  Gender gaps in achievement and participation in multiple introductory biology classrooms.

Authors:  Sarah L Eddy; Sara E Brownell; Mary Pat Wenderoth
Journal:  CBE Life Sci Educ       Date:  2014       Impact factor: 3.325

4.  Males Under-Estimate Academic Performance of Their Female Peers in Undergraduate Biology Classrooms.

Authors:  Daniel Z Grunspan; Sarah L Eddy; Sara E Brownell; Benjamin L Wiggins; Alison J Crowe; Steven M Goodreau
Journal:  PLoS One       Date:  2016-02-10       Impact factor: 3.240

5.  Pandemic precarity: COVID-19 is exposing and exacerbating inequalities in the American heartland.

Authors:  Brea L Perry; Brian Aronson; Bernice A Pescosolido
Journal:  Proc Natl Acad Sci U S A       Date:  2021-02-23       Impact factor: 11.205

6.  The Classroom Observation Protocol for Undergraduate STEM (COPUS): a new instrument to characterize university STEM classroom practices.

Authors:  Michelle K Smith; Francis H M Jones; Sarah L Gilbert; Carl E Wieman
Journal:  CBE Life Sci Educ       Date:  2013       Impact factor: 3.325

7.  The role of gender in scholarly authorship.

Authors:  Jevin D West; Jennifer Jacquet; Molly M King; Shelley J Correll; Carl T Bergstrom
Journal:  PLoS One       Date:  2013-07-22       Impact factor: 3.240

8.  Meta-analysis of Gender Performance Gaps in Undergraduate Natural Science Courses.

Authors:  Sara Odom; Halle Boso; Scott Bowling; Sara Brownell; Sehoya Cotner; Catherine Creech; Abby Grace Drake; Sarah Eddy; Sheritta Fagbodun; Sadie Hebert; Avis C James; Jan Just; Justin R St Juliana; Michele Shuster; Seth K Thompson; Richard Whittington; Bill D Wills; Alan E Wilson; Kelly R Zamudio; Min Zhong; Cissy J Ballen
Journal:  CBE Life Sci Educ       Date:  2021-09       Impact factor: 3.325

9.  A Pandemic Crash Course: Learning to Teach Equitably in Synchronous Online Classes.

Authors:  Daniel L Reinholz; Amelia Stone-Johnstone; Isabel White; Lorenzo M Sianez; Niral Shah
Journal:  CBE Life Sci Educ       Date:  2020-12       Impact factor: 3.325

10.  Female In-Class Participation and Performance Increase with More Female Peers and/or a Female Instructor in Life Sciences Courses.

Authors:  E G Bailey; R F Greenall; D M Baek; C Morris; N Nelson; T M Quirante; N S Rice; S Rose; K R Williams
Journal:  CBE Life Sci Educ       Date:  2020-09       Impact factor: 3.325

View more
  1 in total

1.  Opportunities and Challenges of Online Instruction and Effective Pedagogy That Blurs the Lines between Online and On-Site Teaching and Learning.

Authors:  Andrew Pike; Sumali Pandey; Carlos C Goller; Jennifer Herzog; Samantha T Parks
Journal:  J Microbiol Biol Educ       Date:  2022-04-11
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.