Literature DB >> 33762813

The Process and Product of Coherence Monitoring in Young Readers: Effects of Reader and Text Characteristics.

Nicola K Currie1, Gillian Francey1, Robert Davies1, Shelley Gray2, Mindy S Bridges3, Maria Adelaida Restrepo2, Marilyn S Thompson4, Margeaux F Ciraolo2, Jinxiang Hu5, Kate Cain1.   

Abstract

We examined sixth graders' detection of inconsistencies in narrative and expository passages, contrasting participants who were monolingual speakers (N = 85) or Spanish-English DLLs (N = 94) when recruited in pre-kindergarten (PK). We recorded self-paced reading times and judgments about whether the text made sense, and took an independent measure of word reading. Main findings were that inconsistency detection was better for narratives, for participants who were monolingual speakers in PK, and for those who were better word readers. When the text processing demands were increased by separating the inconsistent sentence and its premise with filler sentences there was a stronger signal for inconsistency detection during reading for better word readers. Reading patterns differed for texts for which children reported an inconsistency compared to those for which they did not, indicating a failure to adequately monitor for coherence while reading. Our performance measures indicate that narrative and expository texts make different demands on readers.

Entities:  

Year:  2020        PMID: 33762813      PMCID: PMC7986958          DOI: 10.1080/10888438.2020.1831503

Source DB:  PubMed          Journal:  Sci Stud Read        ISSN: 1088-8438


Reading comprehension involves the construction of an integrated and coherent representation of the information presented in the text (Johnson-Laird, 1983; Kintsch, 1998). This representation is updated continuously as the text unfolds and readers integrate successive ideas and concepts into the existing model (Rapp & Kendeou, 2007). Theoretically, monitoring the coherence of a text is critical in the construction of a mental representation of its content. Individuals who evaluate the adequacy of their comprehension will detect when information within the text is hard to integrate into the existing mental model, and may take action such as re-reading or inference making (Gernsbacher, 1990; Kintsch, 1998; Rapp & van den Broek, 2005). We extend previous research on coherence monitoring in young readers by examining the influence of critical reader and text characteristics on both the product and process of this skill. The importance of coherence monitoring (also referred to as comprehension monitoring) to successful reading comprehension is evident from studies examining individual differences in reading comprehension and its development: Coherence monitoring is weak in children with poor reading comprehension (Ehrlich, Remond, & Tardieu, 1999; Oakhill, Hartt, & Samols, 2005) and predictive of concurrent and subsequent reading comprehension between 7 to 12 years, over and above word reading, vocabulary, and grammar (Kim, 2015; Language and Reading Research Consortium [LARRC] & Yeomans-Maldonado, 2017; Oakhill & Cain, 2012). Thus, coherence monitoring is well established as critical for successful reading comprehension. Coherence monitoring is typically assessed using an error detection task, in which participants are presented with materials that include deliberate anomalies, such as nonwords, prior knowledge violations, or internal inconsistencies where two details in the text contradict. Many studies require readers (or listeners) to judge whether or not the material makes sense and to identify information that is not coherent with the whole (Baker, 1984; LARRC & Yeomans-Maldonado, 2017; Oakhill et al., 2005). Such measures, taken after the passage has been presented, capture the quality and coherence of the mental representation – the product of text processing. Fewer studies have examined both the product and the process of coherence monitoring in children – what happens when they encounter a coherence break during reading, particularly for passages (Harris, Kruithof, Terwogt, & Visser, 1981; Helder, Van Leijenhorst, & van den Broek, 2016; Zabrucky & Ratner, 1992). This is surprising because one of the earliest studies of coherence monitoring contrasted product and process measures and highlighted the need to examine both to clarify the locus of difficulty (Harris et al., 1981). Processing measures are interpreted in relation to the online detection of a coherence break, with longer times indicating detection and consequent integration difficulty; product measures are interpreted to reflect what information is encoded into the mental model. Harris et al. (1981) presented 8- and 11-year-olds with short narratives and recorded reading times as participants moved a screen to reveal each new sentence. In two experiments, both age groups took longer to read a sentence that was inconsistent with the passage title relative to a consistent condition. However, on completion of the passage, younger children were less likely to correctly report detection of an inconsistency or identify the anomaly. These findings have been reproduced in computer-based reading time studies. Helder et al. (2016) presented 8 to 9- and 10 to 11-year-old good and poor reading comprehenders with narratives, sentence-by-sentence. In half the materials, a coherence break was created by making the second sentence of the passage inconsistent with a characteristic of the protagonist or situation presented in the target final sentence. All readers read the target sentence more slowly when inconsistent, but younger children and poorer comprehenders were less likely to report a coherence break on passage completion. Zabrucky and Ratner (1992) found a similar pattern of findings for 11–12-year-olds. Comparing findings for the product and process measures, each study concluded that all readers detected inconsistencies when present but only the better comprehenders routinely encoded this information in their mental model. Evaluating the adequacy of comprehension is an important skill for reading comprehension in general, and may become more critical in the later grades when reading to learn from text. We examine the influence of reader and text characteristics on both product and process measures to gain a better understanding of the source of coherence monitoring difficulties. The majority of research on coherence monitoring has focused on monolingual speakers. Here, we contrast the performance of monolingual English speakers with children who were identified in PK as Spanish-English dual-language learners (DLLs) whose home language was Spanish, but who were schooled in English. This latter group comprises a sizable population in the U.S. and often shows a poor reading comprehension profile presenting age-appropriate word reading but, by nine years of age, reading comprehension and oral language that lags normative samples even when supported by bilingual education (Lesaux, Crosson, Kieffer, & Pierce, 2010; Nakamoto, Lindsey, & Manis, 2007). Two studies have investigated whether coherence monitoring is a source of DLLs’ weak reading comprehension (Denton et al., 2015; Lesaux & Harris, 2017). Lesaux and Harris (2017) found that 11–13-year-old language minority students reported regular engagement in self-monitoring and in strategic processing to extract meaning from text but demonstrated limited awareness of their comprehension difficulties. As the authors note, these findings contrast with those from a study using a think-aloud procedure with readers aged 12–16, comprising 50% Hispanic students (Denton et al., 2015); reading comprehension difficulties were associated with less frequent use of comprehension processes such as coherence monitoring. Denton et al. (2015) did not report the proportion of Hispanic students in their poor reader sample, limiting comparison between the two studies, and therefore understanding of the reading strengths and weaknesses of DLLs. Our study provides an important extension to this work by contrasting in children identified as monolingual or DLL in PK both the process of coherence monitoring and its product. Text as well as reader characteristics may influence coherence monitoring (Rapp & van den Broek, 2005). Here, we consider the processing demands of the task and genre. Poor comprehenders are less likely to detect inconsistencies when a text has high processing demands involving integration of information across several sentences (Oakhill et al., 2005; van der Schoot, Reijntjes, & van Lieshout, 2012). van der Schoot et al. (2012) found that 10–12-year-old poor comprehenders took longer to read an inconsistent sentence when separated from its predicate by one sentence (low processing demands), but not when there were 5–7 sentences between the two (high processing demands), whilst good comprehenders took longer to read the target sentences in both conditions. The authors concluded that, in contrast to good comprehenders, poor comprehenders do not routinely encode all information into their mental model, so that they only detected inconsistencies when both pieces of information were still active in working memory. Building on this, we examine how the distance between critical information in the text affects both the product and process of coherence monitoring. Few studies of coherence monitoring have contrasted performance between narrative and expository texts (Denton et al., 2015; Zabrucky & Ratner, 1992). Expository texts are considered more challenging than narratives because they can take a variety of structures and contain new information, alongside specialized vocabulary, placing a higher demand on the integration of information within the text and with prior knowledge to support learning (Best, Floyd, & McNamara, 2008; Graesser, McNamara, & Louwerse, 2003). In a recent brain imaging study, 8–10- year-olds made greater use of top-down regions, believed to support the strategic processing associated with coherence monitoring and integration, when reading expository compared to narrative text (Aboud, Bailey, Del Tufo, Barquero, & Cutting, 2019). In addition, Denton et al. (2015) found less evidence of integration, monitoring, and mental model building in think-alouds for expository compared to narrative texts (see also Zabrucky & Ratner, 1992). They suggest that readers in this age group might not go beyond a basic textbase level (or locally coherent) representation for expository texts, indicating that standards of coherence might differ by genre (van den Broek, Risden, & Husebye-Hartman, 1995). If so, readers might be less accurate in detecting inconsistencies in expository texts relative to narratives, but equally likely to accept consistent texts as coherent in both genres. We are the first to compare processing and product measures of coherence monitoring for narrative and expository texts, to determine the locus of difficulty for each genre.

Current study

Our study extends previous research on young readers’ coherence monitoring in several important ways. We examined performance on narrative and expository texts in two groups found to differ in reading comprehension: monolingual English speakers and Spanish-English DLLs. Unlike much previous research, our samples were identified in PK and experienced English-language instruction throughout schooling. Despite this, the evidence that monolingual and DLL children’s reading comprehension skills diverge by around nine years led us to predict that the DLL group would perform more poorly, on average. Texts were either fully consistent or included an inconsistency between two sentences. We manipulated the distance between these two sentences to create two conditions contrasting lower and higher processing demands. We recorded sentence reading times, as well as responses to a sense judgment question after each text. We predicted that coherence monitoring would be poorer for expository relative to narrative texts, and for inconsistent texts with higher processing demands. Our design enables us to determine how differences in language group, genre or processing demands modulate the locus of difficulty detection or encoding of coherence breaks. We also included a measure of children’s word reading, which we predicted would relate to task performance (Perfetti & Stafura, 2014). The close relation of our study design to that used in previous work on monolingual readers’ coherence monitoring, for narrative passages (e.g., Helder et al., 2016; van der Schoot et al., 2012), enables us to test the reproducibility of those study findings, and further to assess the generalizability of their theoretical accounts for different reader groups and genres.

Method

Participants

Eighty-five monolingual English speakers and 94 children who entered pre-kindergarten (PK) as Spanish-English DLLs participated in this study when in grade six (Spring, 2018). Participants were originally enrolled in a 5-year multi-site longitudinal study investigating the language bases of reading comprehension from PK (~4 years) to third grade (~9 years) (for details, see (LARRC, Farquharson, & Murphy, 2016) and reconsented in Grade 6 for the current study. The DLL sample resided in Arizona; the monolingual sample in Arizona, Kansas, and Nebraska. Because the samples were recruited in PK, they attended a number of different schools. Lower income levels were more predominant in the DLL sample (see Table 1). The study confirms to the US Federal Policy for the Protection of Human Subjects and was approved by the Institutional Review Boards or Research Ethics Committees at each university. Informed consent was provided by legal guardians and children gave their assent prior to participation.
Table 1.

Demographic characteristics of english monolingual and Spanish-English dual language learners.

English monolingual speakersSpanish-English Dual Language Learners

N (%female)85 (44%)94 (57%)
Age12 years 1 month12 years 1 month
SWE (raw scores)*78.48 (10.50)75.32 (9.79)
SWE (standard scores)103.49 (15.60)99.06 (13.42)
Income*< 20 k036
20, 001– 40 k1045
40,001– 60 k98
60, 0001– 80 k104
> 80 k551
Father/Male Guardian’s Education Level< High school251
High school1024
Some college123
Associates/Technical degree71
Bachelor’s degree263
Post graduate degree263
Mother/Female Guardian’s Education Level< High school060
High school318
Some college174
Associates/Technical degree84
Bachelor’s degree255
Post graduate degree312
Free/reduced lunch1285

1 non-responder to SWE and Income. SWE refers to the Sight Word Efficiency subtest of the TOWRE-2 (Torgesen, Wagner, & Rashotte, 1999).

Measures and procedure

Children were tested individually in a quiet room in their school or at a university lab.

Sight word reading

Children completed the Sight Word Efficiency subtest of the Test of Word Reading Efficiency – Second Edition (TOWRE-2; Torgesen, Wagner, & Rashotte, 1999), which measures the number of English words, ranging from high to low frequency, pronounced correctly in 45 seconds. The test was administered and scored by trained assessors in line with the manual. The average test-retest reliability reported in the manual is.93.

Coherence monitoring

Each child read 18 eight-sentence narrative texts and 18 eight-sentence expository texts written for this age group. The narratives concerned human characters and events focused on typical activities for children, such as parties, schools, and friendships. The expository texts focused on facts about a specific animal and its habitat. The texts were piloted with 11–12-year-olds using a paper and pen task (where children underlined inconsistencies) to check that target inconsistencies were detected. The administration procedure was piloted with a different sample to check that task instructions and feedback during practice items were clear. Coh-metrix text analysis (Graesser, McNamara, & Kulikowich, 2011)[1] confirmed that the two sets of texts did not differ on word concreteness (Mnarrative = 94.83 (SD = 6.88); Mexpository = 95.61 (SD = 8.12); t < 1.00) but did differ on narrativity (Mnarrative = 50.00 (SD = 18.57); Mexpository = 22.00 (16.50); t(34) = 4.78, p < .001). Within genre, each text was either fully consistent (n = 6) or contained two sentences with contradictory information (n = 12), based on materials from previous studies (e.g., LARRC & Yeomans-Maldonado, 2017; Oakhill et al., 2005). There were two versions for each inconsistent text: in one, the critical sentences were separated by 1–2 sentences (near condition, low processing demands); in the other, they were separated by 3–5 sentences (far condition, high processing demands). The inconsistent items were counterbalanced across two presentation lists to ensure that each participant read only one version of each inconsistent text (to avoid priming) and only completed six passages in each condition. The same six consistent passages were used in both lists. Examples are provided in Table 2. The two sets of texts were equated for length (Mnarrative = 100.66 (SD = 10.32), Mexpository = 103.94 (SD = 8.79), t(34) = 1.03, p = .31) and did not differ in the number of intervening words in the near vs. far inconsistent conditions (Mnarrative = 27.33 (SD = 9.79), Mexpository = 23.17 (SD = 7.28), t(22) = 1.18, p = .25).
Table 2.

Examples of narrative and expository passages in the inconsistent near and far conditions and the consistent condition.

Narrative Inconsistent-NearExpository Inconsistent-Near

Sarah got some roller skates for her birthday.She had never skated before and was surprised at how fast she could skate along the sidewalk.Dad had warned her that she must be very careful not to go too fast, until she got the hang of it.All of a sudden Sarah fell over and very badly hurt her arm.Dad took her to the hospital to get checked by a doctor.The doctor took an X-Ray of Sarah’s leg.The hospital was very busy and they had a long wait ahead of them.Dad promised he would buy Sarah an ice cream on the way home to cheer her up.The monarch butterfly is America’s most familiar butterfly.Its wings have a recognizable black, orange, and white pattern.Monarch butterflies flap their wings more slowly than any other butterfly.They migrate up to three thousand miles each fall.They then fly back again in the spring, traveling up to 350 miles a day.They can make this incredible journey because they can flap their wings more quickly than any other butterfly.The monarch is the only butterfly known to make a two-way migration as birds do.Adult monarch butterflies feed off the nectar of wildflowers and the blossom on fruit trees.
Narrative Inconsistent-FarExpository Inconsistent-Far
Sarah got some roller skates for her birthday.She had never skated before and was surprised at how fast she could skate along the sidewalk.All of a sudden Sarah fell over and very badly hurt her arm.Dad had warned her that she must be very careful not to go too fast, until she got the hang of it.Dad took her to the hospital to get checked by a doctor.The hospital was very busy and they had a long wait ahead of them.The doctor took an X-Ray of Sarah’s leg.Dad promised he would buy Sarah an ice cream on the way home to cheer her up.SENSE QUESTION: Did this story make sense? NOCOMPREHENSION QUESTION: Did Sarah and her dad have to wait at the hospital? YESThe monarch butterfly is America’s most familiar butterfly.Its wings have a recognizable black, orange, and white pattern.Monarch butterflies flap their wings more slowly than any other butterfly.They migrate up to three thousand miles each fall.They then fly back again in the spring, traveling up to 350 miles a day.The monarch is the only butterfly known to make a two-way migration as birds do.They can make this incredible journey because they can flap their wings more quickly than any other butterfly.Adult monarch butterflies feed off the nectar of wildflowers and the blossom on fruit trees.SENSE QUESTION: Does this passage make sense? NOCOMPREHENSION QUESTION: Can monarch butterflies fly more than 500 miles a day? NO
Narrative ConsistentExpository Consistent
Olivia always gets up early in the morning to get ready for school.She often helps to get her little brother Liam ready for school too.This morning she tied her brother’s shoelaces and combed his messy hair.Their mom was busy filling their water bottles and getting the lunch boxes ready.Eventually they were all ready to leave the house.Olivia skipped down the garden path and jumped over the little wall at the end of the garden.Liam is only five years old and very short.He always tries to copy her and ends up falling over the wall instead.SENSE QUESTION: Did this story make sense? YESCOMPREHENSION QUESTION: Did Liam comb his own hair? NOTortoises are land-dwelling reptiles, with hard protective shells.Because they are reptiles, female tortoises lay eggs.Female tortoises do not sit on their eggs like a bird.Instead, they lay the eggs in a burrow and cover them with sand and soil to stay warm.Tortoises are the longest living land animal in the world.Some species of tortoise live for more than 150 years.Their age can be estimated from the rings on the pattern on their shell.The rings can be counted in the same way that we count rings on a tree to estimate its age.SENSE QUESTION: Does this passage make sense? YESCOMPREHENSION QUESTION: Do tortoises sit on their eggs? NO

Bold italicized text indicates inconsistent information in the inconsistent passages. The same consistent passages were used in both lists.

Theoretical interest lies in the difference between the response to sentences in consistent compared to inconsistent near/far conditions. In seeking to estimate the impact of variation in the conditions under which sentences are presented, at the design phase researchers in our field are faced with three options. We could (a) present the same target sentences under all conditions to every participant; (b) present the same target sentences under different conditions to different (sub-groups of) participants; or (c) present different target sentences under different conditions to every participant. Option (a) would be analytically helpful because it could be assumed that any difference in response could not be attributed to differences between-stimuli or between- participants. However, differences between conditions would be confounded with differences in stimulus repetition or order of presentation. Option (b) precludes the risk of confounding the difference between conditions with differences between stimuli but it permits the risk of confounding the difference between conditions with differences between participants or, under counterbalanced designs, between sub-groups of participants. Option (c) precludes the risk of confounding the difference between conditions with differences between participants but it permits the risk of confounding the difference between conditions with differences between stimuli. We chose option (c) because we expected that the inferential risks associated with estimating the effects of conditions in the context of differences between participants would be more important than the risks associated with estimating the differences between conditions in the context of differences between stimuli. Hence, the critical consistent sentences were different from the inconsistent sentences but were presented to elicit responses from the same individuals. (As will be seen, the analysis approach we undertook to examine responses – mixed-effects models – allows us to directly verify our assumptions about the relative balance of inferential risks.) The coherence monitoring task was run using E-Prime 3.0 (Psychology Software Tools, 2016). Instructions outlined the procedure and included an example of an inconsistency. Children completed two practice passages with feedback to ensure their interpretation of the sense question was focussed on the detection of an inconsistency in the text. They viewed the texts on a laptop, advancing to each new sentence by pressing a key on E-Prime’s SR-BOX button box. The reading time for each sentence was recorded. After each text, participants answered a yes/no sense question “Did this text make sense?” (for inconsistent passages a correct response was “no”; for consistent passages a correct response was “yes”). They also answered a comprehension question to encourage reading for meaning (87% correct responses for monolinguals and 80% for DLLs). Each task took approximately 20 minutes to complete. One child did not complete the narrative task and three did not complete the expository task, however, their partial data were included.

Overview of data analysis

Reading time and sense question accuracy data were analyzed with (generalized) linear mixed-effects models (GLMMs) using the lme4 package for R (Bates, Maechler, Bolker, & Walker, 2015). Models were fitted to estimate the predicted (fixed) effects of critical variables (language status, word reading ability, genre and condition) and their interactions while taking into account random effects associated with differences between sampled children or texts. Categorical variables were contrast coded. Word reading scores were standardized. Models specified with maximal random effects structure (Barr, Levy, Scheepers, & Tily, 2013) did not converge so we report models which did converge and which include all fixed effects plus those random effects that were supported by the data (Matuschek, Kliegl, Vasishth, Baayen, & Bates, 2017).

Results

We pre-registered our data preparation and analysis plans (https://osf.io/r69ae), reporting any deviations from our plans in the following. We share our data and analysis code through OSF (https://osf.io/sj28g/).

Responses to sense questions (product of comprehension)

We fitted a GLMM to estimate the effects influencing the log odds that a child’s response to the sense question would be correct, estimating the effects of language status, word reading ability, genre and condition, along with the effects of the interactions between these factors. The mean proportions of correct responses to the sense question in each condition are reported in Table 3. Our model included random effects corresponding to by-items deviations in intercepts, and by-participant deviations in intercepts, and in the slopes of the genre and condition effects. The model summary is reported in Table 4. The positive intercept coefficient shows that participants were more likely to answer the sense question correctly than incorrectly. Given the contrast coding of conditions, the significant positive language status effect (coefficient B = 0.29) shows that monolingual children were 6% more accurate and DLLs were 6% less accurate than the grand mean of the sample.[2] The significant positive genre effect indicates that sense judgments following narratives were more accurate than those following expository texts. The significant negative coefficients representing the effects of differences between the consistent versus the near or far inconsistent conditions show that sense judgments were less accurate for texts containing inconsistencies. The significant positive word reading effect shows that better word reading was associated with more accurate sense judgments.
Table 3.

Mean proportion of correct responses (and Standard Deviations) for the sense question.

Language Status
GenreConditionMonolingualDLLTotal

NarrativeConsistent0.91 (0.29)0.80 (0.40)
Near0.70 (0.46)0.60 (0.49)
Far0.72 (0.46)0.58 (0.49)0.72 (0.45)
ExpositoryConsistent0.90 (0.30)0.75 (0.43)
Near0.41 (0.49)0.39 (0.49)
Far0.42 (0.49)0.39 (0.49)0.54 (0.50)
Total0.68 (0.47)0.59 (0.49)

“Yes” and “no” are the correct responses for consistent and inconsistent texts, respectively.

Table 4.

Summary GLMM for (log odds) sense question accuracy.

Fixed effectsEstimated coefficientSEzp

(Intercept)0.800.126.61<.001
Language status0.290.065.06<.001
Word reading0.140.062.40.02
Genre0.050.114.40<.001
Condition (Near)−0.660.10−6.86<.001
Condition (Far)−0.650.10−6.79<.001
Language status x Genre0.080.042.02.04
Word reading x Genre−0.020.04−0.50.62
Language status x Condition (Near)−0.130.06−2.15.03
Language status x Condition (Far)−0.080.06−1.33.19
Word reading x Condition (Near)−0.070.06−1.10.27
Word reading x Condition (Far)−0.070.06−1.13.26
Genre x Condition (Near)0.190.092.17.03
Genre x Condition (Far)0.160.091.87.06
Language status x Genre x Condition (Near)0.050.041.20.23
Language status x Genre x Condition (Far)0.080.041.76.08
Word reading x Genre x Condition (Near)−0.030.04−0.71.48
Word reading x Genre x Condition (Far)−0.040.04−0.82.41
Random effectsVarianceSD
Participant(intercept)1.071.03
Genre0.240.49
Condition2.431.56
Item(intercept)0.400.63
R2 marginal[c] = 0.22, R2 conditional[d] = 0.45

Observations = 6336[a]; Participants = 178[b]; Items = 36.

One participant had missing narrative data and 3 had missing expository data.

One Monolingual child did not have a TOWRE score. R2 calculated using the MuMIn package in R,

represents the variance explained by the fixed effects,

represents the variance explained by the entire model including both fixed and random effects. Effects in bold are statistically significant. All categorical fixed effects were contrast coded in order to be able to interpret the lower order (main) effects. Language status: Monolingual = +1, DLL = −1; Genre: Narrative = +1, Expository = −1; Condition: Near = +1, Far = +1, Consistent = −1. TOWRE scores were centered and scaled. See Appendix A for the model specification in R and in standard notation.

These main effects were qualified by three significant interactions: language status x genre, language status x condition (near), and genre x condition (near). The nature of these interactions is clearly revealed in Figure 1.
Figure 1.

Graphical representations of the significant interactions in the sense question accuracy model (Table 4).

Traditionally, interaction effects have been explored by sub-setting data to examine the effect of one factor (e.g., language) separately at each level of another factor (e.g., genre). There are important concerns about this approach (see Von der Malsburg & Angele, 2017, for a relevant discussion) that render significance tests problematic but the coefficients estimates from such sub-set analyses are helpful as descriptions of the average differences between conditions (or groups) in outcomes. Thus, in the following, we report estimates but not p-values. Examination of the genre x language status interaction suggests that monolingual participants were more accurate than DLLs for both genres but that the difference due to language status was greater for narrative compared to expository texts (Figure 1a). If we estimate the effect of language status separately for each genre, we see that it is larger for narrative texts (coefficient B = 0.37 (SE = 0.07)) than for expository texts (B = 0.22 (SE = 0.06)). Analysis of the language status x condition interaction (Figure 1b) showed that monolingual children’s sense judgments were correct more often than those of DLL children, for consistent (language status effect, B = 0.52 (SE = 0.10)) and far inconsistent (B = 0.21 (SE = 0.08)) but not for near inconsistent texts (B = 0.16 (SE = 0.09)). Analysis of the genre x condition interaction (Figure 1c) showed that sense questions following inconsistent texts were answered accurately more often for narrative than for expository texts in both the near (genre effect, B = 0.69 (SE = 0.15)) and far (B = 0.65 (SE = 0.16)) but not the consistent condition (B = 0.02 (SE = 0.16)). (See OSF https://osf.io/sj28g/for interaction model summaries.)

Sentence reading times (process of comprehension)

Our analysis of process focused on the comparison of critical sentence reading times for different sentences read by the same children under different conditions. As explained, we assumed that controlling for between-participant differences was more important than controlling for between- stimulus differences. In our pre-registered design, we had planned to estimate the effect of consistency by comparing critical sentence reading times in consistent, near inconsistent, and far inconsistent passages, congruent with the accuracy analysis. (This analysis is reported on OSF https://osf.io/sj28g/). However, we reasoned that while between-stimulus differences were not as important we should still seek to minimize them. Analyzing the effect of condition by comparing sentence reading times in between-passage comparisons confounds condition differences with passage differences. Also, critical sentences varied in length, another potential confound. We resolved both problems as follows. We compared the reading time of each critical sentence in the inconsistent near or far conditions with the time taken to read the sentence located immediately prior (n-1) to the critical sentence in the same text. The n-1 sentence is assumed to have been processed under consistent text conditions because it occurs prior to the critical inconsistent sentence. The comparison of critical and n-1 sentences is within-passage thus removing the confound between passage differences and condition differences. Reading times were scaled to millisecond per word times to remove the potential confound with differences in sentence length. All times were within ±3 SDs of an individual’s condition means. Condition means are reported in Table 5.
Table 5.

Critical sentence reading times (Milliseconds per Word).

Language Status
Text TypeConditionMonolingualDLL

NarrativeInconsistent Near318.91 (151.35)410.01 (230.55)
Inconsistent Far317.73 (148.62)397.73 (196.42)
Consistent Near289.54 (143.07)378.84 (190.66)
Consistent Far284.10 (150.27)380.35 (235.76)
ExpositoryInconsistent Near366.85 (208.25)438.60 (252.32)
Inconsistent Far360.94 (215.47)426.58 (202.66)
Consistent Near356.21 (212.10)436.19 (214.55)
Consistent Far346.75 (200.36)431.64 (276.24)

Reading times for within-text consistent, inconsistent near and inconsistent far conditions

The process model included the same fixed effects as the accuracy model plus sentence type (consistent, inconsistent). It included random effects accounting for between-participant and between- sentence differences in intercepts, between-participant differences in the effects of genre, condition and sentence type, and between-item differences in the effect of word reading. The coefficients show that faster reading times were associated with monolingual status, narrative texts, and better word reading (Table 6); consistent sentences were read more quickly than inconsistent sentences.
Table 6.

Summary LMM for critical sentence reading time (Milliseconds per Word): within texts.

Fixed effectsEstimated coefficientSEtP

(Intercept)372.8910.0537.12<.001
Language status−28.757.29−3.94<.001
Word reading−75.037.51−9.99<.001
Genre−24.968.11−3.08.004
Condition (Near)3.171.911.65.10
Sentence type (Consistent)−8.311.79−4.63<.001
Language status x Genre−3.464.14−0.84.40
Word reading x Genre−2.114.50−0.47.64
Language status x Condition (Near)−0.641.94−0.33.74
Word reading x Condition (Near)2.521.941.30.20
Genre x Condition (Near)−0.871.73−0.50.61
Language status x Sentence type (Consistent)−2.461.82−1.36.18
Word reading x Sentence type (Consistent)−0.591.81−0.33.74
Genre x Sentence type (Consistent)−5.431.73−3.14.002
Condition x Sentence type (Consistent)−1.031.73−0.60.55
Language status x Genre x Condition (Near)−0.371.75−0.21.83
Word reading x Genre x Condition (Near)2.021.751.15.25
Language status x Genre x Sentence type1.291.750.74.46
Word reading x Genre x Sentence type−2.221.75−1.27.20
Language status x Condition x Sentence type1.191.750.68.50
Word reading x Condition x Sentence type2.951.751.69.09
Genre x Condition x Sentence type−0.411.73−0.24.81
Language status x Genre x Condition x Sentence type0.251.750.14.89
Word reading x Genre x Condition x Sentence type0.851.750.49.63
Random effectsVarianceSD
Participant(intercept)8573.0392.59
Genre9562.2197.79
Condition478.8721.88
Sentence type159.3112.62
Text(intercept)1177.5534.32
R2 marginal11 = 0.17, R2 conditional = 0.45Word reading76.278.73

Consistent Near and Far are the comparison sentences (n-1) from the inconsistent passages used in the analysis. DLL = dual language learner.

Observations = 8448[a], Participants = 178[b], Texts = 24.

There were two items (consistent/inconsistent) per text. One participant had missing narrative data and 3 had missing expository data.

One Monolingual child did not have a TOWRE score. R2 calculated using the MuMIn package in R,

represents the variance explained by the fixed effects,

represents the variance explained by the entire model including both fixed and random effects. Effects in bold are statistically significant. All categorical fixed effects were contrast coded in order to be able to interpret the lower order (main) effects. Language status: Monolingual = +1, DLL = −1; Genre: Narrative = +1, Expository = −1; Condition: Near = +1, Far = −1, Sentence type: Inconsistent = +1, Consistent = −1. TOWRE scores were centered and scaled. See Appendix B for the model specification in R and in standard notation.

The genre sentence type interaction was statistically significant (illustrated in Figure 2). Analyses of the effect of sentence type, considered separately for narrative and expository texts, indicate that consistent sentences were read more quickly than inconsistent sentences for narratives (B = −13.74 (SE = 2.09)) but not for expository texts (B = −2.87 (SE = 2.78); see OSF https://osf.io/sj28g/for interaction model summaries).
Figure 2.

Graphical representation of the significant genre x sentence type interaction in the all data sentence reading time model (Table 6).

We assumed that inferential risks were greater in comparing conditions if manipulated between- participants than if manipulated between-stimuli. We can examine this assumption directly. The random effects variances of our analyses (see Tables 6–8) are estimates but, the variances associated with random differences between participants are considerably larger than the random effects variances associated with random differences between stimuli. This suggests it is preferable, given available options, to compare responses to different sentences under different conditions for the same participants.
Table 8.

Summary LMM for narrative text reading time (Milliseconds per Word): incorrect responses.

Fixed effectsEstimated coefficientSEtP

(Intercept)369.8310.5235.15<.001
Language status−32.498.16−3.98<.001
Word reading−79.818.21−9.72<.001
Genre−23.428.16−2.87.007
Condition2.573.060.84.40
Sentence type−2.372.82−0.84.40
Language status x Genre−9.344.64−2.01.05
Word reading x Genre−6.624.72−1.40.16
Language Status x Condition−1.093.09−0.35.72
Word reading x Condition (Near)3.873.181.22.22
Genre x Condition (Near)−0.882.95−0.30.77
Language status x Sentence type (Consistent)−1.822.83−0.65.52
Word reading x Sentence type (Consistent)2.092.930.72.47
Genre x Sentence type (Consistent)−2.342.82−0.83.41
Condition x Sentence type (Consistent)4.392.821.56.12
Language status x Genre x Condition−0.682.97−0.23.82
Word reading x Genre x Condition2.883.070.94.35
Language status x Genre x Sentence type3.272.831.16.25
Word reading x Genre x Sentence type−0.952.93−0.33.74
Language status x Condition x Sentence type3.362.831.19.23
Word reading x Condition x Sentence type1.272.930.44.66
Genre x Condition x Sentence type1.102.820.39.70
Language status x Genre x Condition x Sentence type−0.612.83−0.22.83
Word reading x Genre x Condition x Sentence type0.232.930.08.94
Random effectsVarianceSD
Participant(intercept)8242.0090.79
Genre7814.7088.40
Condition461.2021.47
Text(intercept)1063.8032.62
R2 marginal11 = 0.15, R2 conditional = 0.42

Observations = 3998[a], Participants = 178[b], Texts = 24.

There were two items (consistent/inconsistent) per text.

One Monolingual child did not have a TOWRE score. R2 calculated using the MuMIn package in R,

represents the variance explained by the fixed effects,

represents the variance explained by the entire model including both fixed and random effects. Effects in bold are statistically significant. All categorical fixed effects were contrast coded in order to be able to interpret the lower order (main) effects. Language status: Monolingual = +1, DLL = −1; Genre: Narrative = +1, Expository = −1; Condition: Near = +1, Far = −1, Sentence type: Inconsistent = +1, Consistent = −1. TOWRE scores were centered and scaled. See Appendix B for the model specification in R and in standard notation.

Critical sentence reading times in correct and incorrect responses

Similar to Helder et al. (2016), we examined whether the pattern of effects differed when children answered correctly or incorrectly. Sentence reading times were analyzed separately conditional on sense question response accuracy. The coefficients show that, given correct sense judgments, just as in the analysis of all responses, faster reading times were associated with monolingual status, narrative texts, better word reading and consistent sentences (Table 7). There was a genre x sentence type interaction (See Figure 3) where sub- set analyses indicated that consistent sentences were read more quickly than inconsistent sentences in narrative (B = −18.21 (SE = 2.54)) but not expository texts (B = −7.07 (SE = 4.12)). There was a significant condition x sentence type interaction, qualified by a three-way interaction with word reading ability (See Figure 3). Sub-set analyses indicated that in the near condition, word reading ability did not modulate the difference in reading times between consistent and inconsistent sentences (B = 2.41 (SE = 3.35)) while, in the far condition, better word reading was associated with a greater difference in reading times between the two sentences types (B = −9.68 (SE = 3.62); see OSF https://osf.io/sj28g/for model summaries).
Table 7.

Summary LMM for reading time (Milliseconds per Word): Correct responses.

Fixed effectsEstimated coefficientSEtP

(Intercept)374.8610.4935.73<.001
Language status−28.577.50−3.81<.001
Word reading−73.007.41−9.85<.001
Genre−25.878.90−2.91.006
Condition1.912.480.77.44
Sentence type−12.642.29−5.52<.001
Language status x Genre−3.224.94−0.65.51
Word reading x Genre−3.064.83−0.63.52
Language Status x Condition0.082.520.03.98
Word reading x Condition (Near)0.672.460.27.78
Genre x Condition (Near)0.602.380.25.80
Language status x Sentence type (Consistent)−1.442.34−0.62.54
Word reading x Sentence type (Consistent)−2.442.28−1.07.29
Genre x Sentence type (Consistent)−5.572.29−2.43.01
Condition x Sentence type (Consistent)−5.792.29−2.53.01
Language status x Genre x Condition−0.252.42−0.11.92
Word reading x Genre x Condition2.492.361.06.29
Language status x Genre x Sentence type0.072.340.03.98
Word reading x Genre x Sentence type−2.052.28−0.90.37
Language status x Condition x Sentence type−1.212.34−0.52.61
Word reading x Condition x Sentence type4.852.282.13.03
Genre x Condition x Sentence type0.722.290.31.75
Language status x Genre x Condition x Sentence type2.552.341.09.27
Word reading x Genre x Condition x Sentence type0.132.280.06.95
Random effectsVarianceSD
Participant(intercept)8676.7093.15
Genre11,036.10105.05
Condition316.3017.79
Text(intercept)1320.7036.34
R2 marginal11 = 0.20, R2 conditional = 0.48

Observations = 4450[a], Participants =175[b], Texts = 24.

There were two items (consistent/inconsistent) per text.

One Monolingual child did not have a TOWRE score. Three participants did not respond correctly to any of the items. R2 calculated using the MuMIn package in R,

represents the variance explained by the fixed effects,

represents the variance explained by the entire model including both fixed and random effects. Effects in bold are statistically significant. All categorical fixed effects were contrast coded in order to be able to interpret the lower order (main) effects. Language status: Monolingual = +1, DLL = −1; Genre: Narrative = +1, Expository = −1; Condition: Near = +1, Far = −1, Sentence type: Inconsistent = +1, Consistent = −1. TOWRE scores were centered and scaled. See Appendix B for the model specification in R and in standard notation.

Figure 3.

Graphical representation of the significant interactions in the correct only sentence reading time model (Table 7).

In comparison, given incorrect sense judgments, faster reading times were associated with monolingual status, narrative texts, and better word reading (Table 8) but the effect of sentence type was not significant. There was a significant language status x genre interaction (Figure 4) which, in sub-set analyses, appeared because genre had a significant effect for monolingual (B = −30.66 (SE = 7.72)) but not DLL participants (B = −15.23 (SE = 9.97); see https://osf.io/sj28g/for model summaries).
Figure 4.

Graphical representation of the significant language status x genre interaction in the incorrect only sentence reading time model (Table 8).

Discussion

Our examination of 11–12-year-olds’ reading of consistent and inconsistent texts provides unique insight into how reader and text attributes influence coherence monitoring. We extend previous research on coherence monitoring by demonstrating that readers are more likely to detect and report an inconsistency if they are monolingual speakers compared to Spanish-English DLLs (when recruited in PK), if they are better word readers, and when passage content is narrative. The processing demands of the task differentially influenced the strength of the signal that a coherence break was present for good and poor word readers while, overall, participants took longer to read inconsistent sentences only for passages eliciting accurate sense judgments. Our findings suggest that the primary locus of coherence monitoring failure lay in not detecting an inconsistency while reading, rather than not encoding this information. Similar to previous reading time studies of passage-level coherence monitoring (Harris et al., 1981; Helder et al., 2016; van der Schoot et al., 2012; Zabrucky & Ratner, 1992), children differentiated between consistent and inconsistent text. They took longer to read inconsistent sentences and reported these coherence breaks on completion of the passage. (Agreement on consistent texts was high across genre and language status groups.) Critically, we found reading time differences between consistent and inconsistent sentences only for passages eliciting accurate sense judgments. Thus, we propose that when a coherence break was not reported (incorrect sense judgment after reading) it was not detected during reading. This contrasts with proposals that coherence monitoring failures arise when children detect, but do not encode, a coherence break in their mental model. Most previous studies of the process of coherence monitoring have used only narrative texts (Harris et al., 1981; Helder et al., 2016; van der Schoot et al., 2012). We found that genre matters, with lower levels of detection during reading and subsequent reporting of coherence breaks for inconsistent expository texts. This pattern suggests that readers are more likely to process expository text in a piecemeal sentence-by-sentence manner, perhaps adopting different standards of coherence for different genres (van den Broek et al., 1995), and thus failing to integrate information across sentences to construct a coherent mental model of the whole text (Denton et al., 2015). We found longer reading times for expository texts in general, further indicating they were more challenging than the narratives (Best et al., 2008; Graesser et al., 2003). Given recent imaging research showing that narrative and expository text make different processing demands (Aboud et al., 2019), future research should examine how the genre-related differences in standards of coherence or processing that we observe are related to genre differences in content (structure, vocabulary) and reading goals (learning vs. pleasure) (Graesser et al., 2003). Imaging studies using temporal measures could further elucidate the locus of difficulty for different genres. DLLs were less likely than monolinguals to monitor text for coherence, but the locus of difficulty for both groups appeared to be the same: a failure to detect inconsistencies when reading. The DLLs in our study were not selected (at PK) to be poor comprehenders but previous research, consistent with our findings, suggests that many may have a poor comprehender profile (Lesaux et al., 2010; Mancilla- Martinez & Lesaux, 2011; Nakamoto et al., 2007). Contrary to predictions, the manipulation of the processing demands of our texts did not influence detection or reporting of inconsistencies (see also, Zabrucky & Ratner, 1992). However, poorer word readers showed a smaller processing time difference between consistent and inconsistent sentences when the processing demands were high. This suggests that a coherence break is more readily detected by stronger (compared to weaker) readers under high demand conditions. This pattern mirrors that reported for good and poor comprehenders by van der Schoot et al. (2012). Our study has several important strengths, including the comparison of product and process measures, of genre, and of language groups, besides our use of mixed-effects models to account for random differences. We discuss the limitations here. First, like other work in this field, our groups differed in relation to both language background and socio-economic status. Thus, despite English language schooling from PK, our DLLs may have shown poorer coherence monitoring because of unobserved effects of low-income and language exposure at home (Hoff, 2013). Isolation of the influence of each factor is needed to inform targeted support. Second, we included both product and process measures to permit identification of the most likely source of coherence monitoring failures. Through this approach, our readers were necessarily alerted to the presence of inconsistencies and may have adopted different standards of coherence and strategies compared to “typical” reading. Examining the effect of task instructions on the process of reading could provide important information about the influence of different reading goals, and insight into ways to foster better coherence monitoring. Relatedly, we varied the position of the inconsistent sentence to minimize strategic processing. We note that the inconsistent sentence in Helder et al.’s (2016) materials was always in the same sentence-final position, which may have encouraged more strategic anticipatory processing with the resultant higher accuracy scores than we report here. Finally, our reading time paradigm did not permit examination of whether readers looked back to check preceding text when an inconsistency was detected. Future studies could use eye tracking methods to do this (Connor et al., 2015) and determine whether this behavior differs by reader or genre (Zabrucky & Ratner, 1992). In summary, we have advanced understanding of young readers’ coherence monitoring, indicating for the first time that this skill is weaker in Spanish-English DLLs. Our findings indicate the most likely source of poor coherence monitoring is a failure to construct a coherent mental model when reading, rather than a failure to encode a break when detected. This is particularly evident for expository texts. Future work should identify how the teaching and activation of relevant vocabulary, background knowledge and reading strategies could support the development of coherence monitoring, a critical skill for learning, and a fundamental skill for expository text comprehension.
  1 in total

1.  A measure of individual differences in readers' approaches to text and its relation to reading experience and reading comprehension.

Authors:  Regina C Calloway; Anne Helder; Charles A Perfetti
Journal:  Behav Res Methods       Date:  2022-05-03
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.