Literature DB >> 33569848

Taking "the boss" into the real world: Field interrater reliability of the Short-Term Assessment of Risk and Treatability: Adolescent Version.

Tamara L F De Beuf1,2, Corine de Ruiter2, John F Edens3, Vivienne de Vogel4.   

Abstract

There is emerging evidence that the performance of risk assessment instruments is weaker when used for clinical decision-making than for research purposes. For instance, research has found lower agreement between evaluators when the risk assessments are conducted during routine practice. We examined the field interrater reliability of the Short-Term Assessment of Risk and Treatability: Adolescent Version (START:AV). Clinicians in a Dutch secure youth care facility completed START:AV assessments as part of the treatment routine. Consistent with previous literature, interrater reliability of the items and total scores was lower than previously reported in non-field studies. Nevertheless, moderate to good interrater reliability was found for final risk judgments on most adverse outcomes. Field studies provide insights into the actual performance of structured risk assessment in real-world settings, exposing factors that affect reliability. This information is relevant for those who wish to implement structured risk assessment with a level of reliability that is defensible considering the high stakes.
© 2021 The Authors. Behavioral Sciences & The Law published by John Wiley & Sons Ltd.

Entities:  

Mesh:

Year:  2021        PMID: 33569848      PMCID: PMC7986435          DOI: 10.1002/bsl.2503

Source DB:  PubMed          Journal:  Behav Sci Law        ISSN: 0735-3936


INTRODUCTION

One of the first steps in implementing structured risk assessment is the selection of an appropriate risk assessment instrument. When deciding on the most suitable instrument for a particular setting or jurisdiction, multiple factors should be considered, such as feasibility, possibility of training, and costs (Vincent et al., 2012). The psychometric properties of potential instruments should also guide the decision, and peer‐reviewed empirical research on the reliability and validity of an instrument should be evaluated. In regard to risk assessment instruments, reliability generally refers to the consistency of results when the assessment is repeated, and validity addresses whether the instrument's scores accurately predict the occurrence of the outcome of interest. Contrary to what the term “properties” suggests, reliability and validity are not intrinsic features of any risk assessment instrument; they are not static characteristics that automatically translate to other settings, populations, and evaluators [DeMatteo et al., 2020; Edens et al., 2013; more generally, see the Standards for Educational and Psychological Testing (American Educational Research Association, American Psychological Association, and the National Council on Measurement in Education, 2014)]. Performance may differ not only by location or setting (e.g., prison vs. community service), but also by the evaluator's assessment goal: data collection for research purposes versus data generated for clinical/legal practices (Edens & Boccaccini, 2017). There is emerging evidence that instruments perform better in more controlled studies (hereafter referred to as “non‐field studies”) than when administered by practitioners in everyday practice (Neal et al., 2015). In recent years, this awareness has contributed to a growing effort to assess the performance of risk assessment instruments in real‐world settings, referred to as field studies (Edens & Boccaccini, 2017).

Field studies of risk assessment

Field studies are distinct from non‐field studies in that data are generated as part of clinical, forensic, or correctional practice. Thus, field studies are based on an assessment outcome “used as part of a decision‐making process that would have real‐world implications for the person being evaluated” (Edens & Boccaccini, 2017, p. 600). Assessments conducted by professionals in the field with the sole purpose of research are not considered field studies by this definition. Field studies on reliability may be particularly important, as there is emerging evidence that lower interrater reliability may limit the predictive validity of risk assessment instruments in real‐world settings (Edens & Kelley, 2017). For example, in a meta‐analysis, Hanson and Morton‐Bourgon (2009) found that risk assessment studies with lower interrater reliability had significantly lower effect sizes for recidivism than studies with higher reliability. Low reliability indicates measurement error that is either nonsystematic (i.e., random) or systematic (e.g., rater bias); the more error, the more difficult it is to make valid predictions (Boccaccini et al., 2009; Neal et al., 2015). For this reason, Edens and Kelley (2017) labeled reliability, and particularly interrater reliability, as “the boss” of psychometric qualities in the field.

Field interrater reliability

Field reliability is often assessed in terms of interrater reliability, that is, the consistency of findings when different evaluators assess the same subject. Establishing interrater reliability is relevant to ensure the fairness of decisions that result from the outcome of the risk assessment. It can be understood as a quality check for those who decide to implement the instrument within their practice. Until now, the majority of studies on field interrater reliability in forensic assessment involved actuarial risk assessment instruments for (sexual) violence (Boccaccini et al., 2009; Edens et al., 2016; Miller et al., 2012; Rettenberger et al., 2017; Schmidt et al., 2005) and the Psychopathy Checklist‐Revised (PCL‐R; Hare, 2003), an instrument to assess psychopathy, also commonly used to aid in risk assessment (Edens et al., 2019). These studies have typically been conducted on test results that were collected specifically to inform case decision‐making as part of some adversarial judicial proceeding (e.g., criminal trial or sexually violent predator civil commitment hearing). However, there is emerging literature on the field interrater reliability of structured professional judgment (SPJ) risk assessment instruments that are frequently implemented in post‐adjudication forensic psychiatric settings (de Vogel & de Ruiter, 2004; Jeandarme et al., 2017; Nicholls et al., 2006; Troquete et al., 2015). Findings suggest that risk assessments completed by practitioners as part of routine decision‐making (e.g., leave status, discharge) – whether using actuarial instruments or SPJ approaches – obtain weaker interrater reliability than assessments in controlled non‐field studies (e.g., Jeandarme et al., 2017). However, some measures seem to be less affected (e.g., Vincent et al., 2012). Authors have presented various arguments for this decrement in interrater reliability. They argue that studies in the field are conducted in less‐than‐ideal conditions, whereas in non‐field studies, conditions can be better controlled. For example, evaluators in field studies are more likely to show wider variability in terms of experience, training, and/or professional background, compared with research assistants in non‐field studies (Boccaccini et al., 2008; de Vogel & de Ruiter, 2004). They are also less likely than research assistants to have received consistent and extensive training, and their risk assessment performance is usually not as closely monitored or supervised (DeMatteo et al., 2020; Guarnera & Murrie, 2017). Even when the evaluators in a non‐field study are practitioners from the field, awareness that they are participating in an empirical study or that researchers are involved, could improve the evaluators' usual performance (Penney et al., 2014; Vincent et al., 2012). When individuals know they are being evaluated and their errors are being monitored, they will perform differently (Dror, 2020). Additionally, risk assessments conducted for clinical purposes might be influenced by contextual pressures, such as political or organizational demands (e.g., to facilitate a transfer of the patient; Jeandarme et al., 2017) and emotional biases (e.g., positive vs. negative emotional involvement with the patient; de Vogel & de Ruiter, 2004). Furthermore, in contrast to non‐field studies in which evaluators usually receive the same information package or case vignette, clinicians in field studies are responsible for gathering their own information (Boccaccini et al., 2008). In practice, information might not only be qualitatively and quantitatively limited compared with more controlled studies, but the information used for the assessment might also differ per evaluator (DeMatteo et al., 2020; Penney et al., 2014). Similarly, it is possible that interviews conducted for research purposes versus clinical‐legal purposes will elicit a different degree of honesty, disclosure, and impression management from the person evaluated, beyond the effect of an evaluator's interview style (Edens & Kelley, 2017; Guarnera & Murrie, 2017). In addition to these evaluator and contextual characteristics, features of the risk assessment instrument itself can also impact reliability. For example, item subjectivity and a lack of clear instructions potentially increase noise in the ratings and may limit agreement between evaluators (Edens & Kelley, 2017). As such, actuarial risk assessment instruments that mainly consist of static factors targeting quantifiable behaviors (e.g., the number of prior charges) are more likely to be completed reliably than instruments that include interpersonal and affective features that require a certain degree of clinical judgment (e.g., self‐esteem, empathy; Edens et al., 2019). Indeed et al. (2011) found moderate to large inverse correlations between the perceived subjectivity of risk factors and interrater agreement on these factors, suggesting that at least some variation in reliability could be explained by item subjectivity. Similarly, Kennealy et al. (2016) found lower interrater agreement for features that required more clinical judgment (e.g., attitudes) compared with more straightforward features (e.g., criminal history). Given the more central role of evaluator judgment in the use of SPJ risk assessment instruments, these instruments might be particularly prone to the above‐mentioned influences when applied in the field, resulting in lower field interrater reliability compared with actuarial instruments (Edens & Kelley, 2017).

Field reliability of the START:AV

A fairly recently developed SPJ instrument for adolescents is the Short‐Term Assessment of Risk and Treatability: Adolescent Version (START:AV; Viljoen et al., 2014). This instrument assesses the short‐term risk of eight adverse outcomes, such as violence to others and victimization (see the Method section). Evaluators reach a final risk judgment (low, moderate, or high) by weighing historical information against the adolescent's current presentation on items related to characteristics of the adolescent (e.g., impulsivity), their relationships (e.g., parenting), and their response to treatment (e.g., insight). The START:AV items are combined into overall ratings of “strengths” and “vulnerability,” and total scores for both can be calculated for research purposes. With half a dozen publications, research on the psychometric properties and utility of the START:AV is still in its infancy (Desmarais et al., 2012; Klimukienė et al., 2018; Sellers et al., 2017; Sher et al., 2017; Singh et al., 2014; Viljoen et al., 2012). Moreover, most of these studies have used the pilot version of the instrument. To our knowledge, there are no START:AV field studies in the literature as yet. In the following, we describe the results of two non‐field studies that have examined the interrater reliability of the START:AV. The first interrater reliability examination was conducted by the instrument's authors and involved 12 START:AV assessments of adolescent offenders on probation completed by graduate students (Viljoen et al., 2012). The students received training on the interview process and the administration of the START:AV and completed multiple practice cases, including an in‐person practice assessment alongside an experienced rater. Before conducting START:AV assessments for data collection, the students had to demonstrate interrater reliability consistent with a consensus or gold standard rating, that is, the ratings agreed by a team of START:AV experts. The student evaluators based their assessments on the youth's justice records and an interview with the adolescent. Intraclass correlation coefficients (ICCs; two‐way random‐effects model, absolute agreement, single measure) for the total scores were 86 for vulnerabilities and 92 for strengths. ICCs for six of the START:AV's final risk judgments were as follows: 52 for suicide, 60 for violence to others and substance abuse, 75 for general reoffending, 86 for victimization and 88 for self‐injury. More recently, the START:AV was evaluated in a Lithuanian juvenile probation sample (Klimukienė et al., 2018). Similar to the study by Viljoen et al. (2012), this was a non‐field study in which the START:AV assessments were conducted by a research team (not further specified). The research team practiced with interviews gathered during a pilot study until they reached agreement on the pilot START:AV assessments. For the actual interrater reliability study, the team randomly selected 30 interviews that involved a team member questioning a juvenile probation officer about an adolescent under their supervision. These audio‐taped interviews were subsequently used as the only source of information for the risk assessments. ICCs (two‐way random‐effects model, absolute agreement, single measure) for the total scores were 91 for vulnerabilities and 82 for strengths. Additionally, ICCs for four of START:AV's final risk judgments were calculated: 49 for substance abuse, 51 for violence to others, 66 for general reoffending, and 89 for unauthorized absences. Klimukienė et al. (2018) also analyzed the interrater reliability of the items. The ICCs for vulnerabilities ranged from 29 to 86 (average ICC = 67), with lowest agreement for external triggers (ICC = 29), parental functioning (ICC = 37), and social support from peers (ICC = 39). For the strengths items, ICCs ranged from 36 to 80 (average ICC = 58), with lowest agreement for coping (ICC = 36), social support from adults (ICC = 39), and community (ICC = 39). The authors argued that the low agreement for these items was likely due to the fact that the interview with the juvenile probation officer was the only source of information.

Present study

Beyond these two non‐field studies with juvenile probation samples, there have been no studies investigating the interrater reliability of the START:AV, nor have there been field studies on this issue. As underlined by Webster et al. (2002) almost two decades ago and recently reiterated by Edens and Boccaccini (2017), there is a need for risk assessment studies using data from clinicians during their everyday work. The present study addresses the need for field studies in relation to the START:AV. This instrument was implemented in a Dutch secure residential youth care facility where the risk assessments were used for decision‐making regarding level of supervision and treatment objectives. In this study, we examine “the boss” of psychometric qualities (i.e., interrater reliability) and in line with the emerging field research, we expect to find lower interrater reliability for the START:AV features than previously reported in the non‐field studies.

METHOD

Setting

The study was conducted in one of 14 secure residential youth care facilities in the Netherlands. Secure residential services provide the most intensive type of youth care and are often considered a “last resort” for boys and girls with severe behavioral and/or mental health problems (Ten Brummelaar et al., 2017). Adolescents can only be admitted to residential youth care by a court order. This setting is distinct from juvenile detention centers in that youth are admitted under civil law rather than criminal law. Mandated youth care occurs under this program when deemed necessary by a judge to guarantee the adolescents' safety (e.g., from self‐harm, neglect) and/or the safety of their environment (e.g., violence to others, criminal activities). At the present location, 137 adolescents were admitted in 2016 and 225 adolescents (52% girls) were in treatment over the course of that year. The average treatment duration was 262 days (8.6 months, range: 4–717 days). At the time of the study, the service had three high secure units (including two observation units) and six medium secure units.

Youth sample

The START:AV forms (n = 100) were randomly selected from all START:AVs completed between 1 March 2016 and 2 November 2017 that had fewer than six missing strength or vulnerability ratings (N = 237). The START:AV assessments concerned 82 unique adolescents; several (20%) had multiple START:AV forms. The sample consisted of 41 male and 41 female adolescents (in line with the setting's boy:girl ratio) between the ages of 12.4 and 17.9 years at the time of the START:AV assessment (M = 16.3, SD = 1.3). They spent, on average, 300 days (9.9 months) in medium and/or high secure care (SD = 131.17; range = 66–664). Forty‐five percent resided in a high secure ward (24% observation, 21% treatment), while the others resided in a medium secure treatment unit. The sample's average total IQ score, measured with the Wechsler Intelligence Scale for Children‐V‐NL (WISC‐V‐NL; Wechsler, 2014/2017), was 87.7 (SD = 14.38) and ranged from 54 to 119. For 28% of the adolescents, no total IQ score could be calculated because of a disharmonic intelligence profile (i.e., 1 SD or more discrepancy between verbal and performance IQ score). All youths had at least one diagnosis according to the Diagnostic and Statistical Manual of Mental Disorders (5th ed.; American Psychiatric Association, 2013). The most common diagnoses were oppositional defiant disorder (n = 22, 33%), conduct disorder (n = 26, 32%), attention deficit hyperactivity disorder (n = 25, 31%), as well as post‐traumatic stress disorder (n = 19, 23%), substance‐related disorders (n = 17, 21%), intellectual disability (n = 16, 20%), and autism spectrum disorder (n = 11, 13.4%). A majority of the adolescents had three (n = 24, 30%) or four diagnoses (n = 26, 32%), and 17% (n = 14) had more than four diagnoses. Non‐selected adolescents did not differ significantly from the included sample concerning these characteristics.

Evaluators

Clinical evaluators

The START:AV forms were completed as part of routine practice by 12 evaluators who were employed as treatment coordinators within the service. These professionals, with at least a master's degree in psychology or special needs education, were responsible for the adolescents' treatment process. In consultation with the team on the unit, they made decisions about treatment (e.g., treatment goals, types of therapy) and evaluated treatment progress. All evaluators were women, aged 26–40 years (M = 32.5, SD = 4.4), with on average 6 years of service as a treatment coordinator within the organization (SD = 4.8, range: 0–14). Prior to the START:AV training, 50% had previously used risk assessment tools in practice and 60% had received some sort of risk assessment training. All evaluators were trained by the first author who, in turn, had received training and supervision from the instrument's developers. Nine evaluators participated in a 2‐day group training event and discussed additional practice cases during another 2‐hour training session led by the first and third authors. Three evaluators, hired after the initial training, received one‐on‐one training shortly after starting employment. In the present sample, the evaluators completed, on average, eight START:AVs (SD = 6), ranging from one to 21 START:AVs per evaluator.

Research evaluators

Four female graduate‐level students in psychology conducted independent ratings on the START:AV as research assistants. They all had completed coursework on forensic populations and adolescent development and had experience with young people in residential youth care. One research evaluator (first author) was trained by the original authors of the START:AV, two evaluators participated in a 1‐day group training event provided by the first author followed by one‐on‐one feedback sessions discussing practice cases, and the fourth research rater was trained one‐on‐one by the first author, including one‐on‐one feedback on multiple practice cases.

Measure

The START:AV is an SPJ risk assessment instrument assessing the short‐term risk of multiple adverse outcomes, including violence to others, nonviolent offending, substance abuse, unauthorized absences, suicide, self‐injury, victimization, and health neglect. Prior to formulating a final risk judgment (low, moderate, or high), the evaluator rates 24 items (low, moderate, or high) based on information of the previous months and assesses the occurrence of adverse outcomes in the recent and prior history (present/absent). All START:AV items are dynamic items that can change over time due to normal development or in response to life events and interventions (Viljoen et al., 2012). The items involve behavioral, cognitive, and interpersonal characteristics of the adolescent and how they respond to interventions, as well as features of their relationships and environment (e.g., “social support,” “peers”). Items are assessed in two directions: as strengths (i.e., positive features that may reduce risk) and as vulnerabilities (i.e., difficulties that may increase risk). The START:AV user guide (Viljoen et al., 2014, 2016) provides a definition for each item as well as indicators (i.e., item anchors) for both strengths and vulnerabilities. In addition to the item anchors, the user guide provides examples of moderate versus high strengths/vulnerabilities. After rating the items, the evaluator identifies which strengths and vulnerabilities are particularly relevant to future risks. Key strengths may be important in protecting against adverse outcomes, whereas critical vulnerabilities could play a crucial role in contributing to adverse outcomes. The key and critical items thus identified are building blocks for the risk formulation and the intervention plan.

Implementation

In February 2016, the service introduced the START:AV as the risk assessment instrument within its practice (see also De Beuf et al., 2020). The START:AV was embedded within the treatment cycle: every 4 months, the START:AV was completed again as part of the periodic treatment plan evaluation. At first, comprehensive rating forms were completed and coded by treatment coordinators. Starting in mid‐2017, group care workers and other care professionals (e.g., therapists, family social workers) also started reporting their observations in the START:AV form, but the treatment coordinators remained responsible for rating the items and adverse outcomes. Various sources of information were used to complete the risk assessments: file review of internal (e.g., daily progress notes, previous START:AVs) and external (e.g., supervision order, treatment reports from previous service agencies) documentation, as well as interviews with the adolescent, their caregivers and/or other collateral informants. In addition, treatment coordinators had the opportunity to observe the adolescent on the unit.

Procedure

The clinician‐rated START:AV forms were gathered from the patient files with permission from the service's general director and approval from the Ethics Review Committee Psychology and Neuroscience (ERCPN) of Maastricht University (ERCPN number 174_05_12_2016). As described earlier, the clinician‐rated forms were completed as part of routine clinical practice. The second rating of each START:AV was completed by a research evaluator; it was never the case that two clinicians rated the same case for clinical purposes. Therefore, research assistants (see section on Research evaluators) conducted the second assessments for 100 randomly selected START:AVs. Although the research evaluators completed the assessment at another time point compared with the clinical evaluators, they relied on the same file information (i.e., internal and external documentation) that was available to the clinical evaluators at the time of their assessment.

Statistical analysis

Before assessing interrater reliability, we explored several features of the sample, including missing ratings and the distribution of scores per evaluator group. When there were missing items, total scores were prorated by dividing the raw total score by the maximum total score, and multiplying this by 52, because the maximum total score for 26 items (items 13 and 14 each have two separate ratings), each with a maximum score of 2, is 52. When item 23 (medication adherence) was rated as “not applicable,” the item was handled as a missing item. The prorating formula differs from previous studies with the START:AV (Desmarais et al., 2012; Klimukienė et al., 2018), but is consistent with other risk assessment instruments (Boer et al., 1997; Webster et al., 1997). We tested the above formula and the formula applied in previous START:AV studies on a subsample of complete START:AVs in which we artificially created missing ratings. We found that the prorating formula of our choice best approximated the true values in the sample. We calculated interrater reliability of prior and recent history of adverse outcomes using Gwet's agreement coefficient (AC1; Gwet, 2002). This statistic is preferred to a kappa coefficient because Gwet's AC is not affected by prevalence rates, whereas kappa tends to underestimate agreement in situations with high or low prevalence (e.g., prevalence of suicide), a phenomenon referred to as the “kappa paradox” or the “paradox of high agreement, low reliability” (Feinstein & Cicchetti, 1990). Gwet's AC1 was calculated using the AgreeStat360 Excel program (Gwet, 2020). For the interrater reliability of the individual items, the strength and vulnerability total scores, and the eight final risk judgments, the ICC was calculated. More specifically, a single‐measures, two‐way random model was selected, using both the consistency and absolute definition (Koo & Li, 2016). A two‐way random ICC was deemed appropriate because the forms were completed by evaluators from the same set of evaluators who represented a larger group of evaluators of interest. A single‐measures type was most relevant to the current risk assessment practice as every assessment was conducted by a single evaluator (“single measures” type) rather than using the average score of multiple evaluators (“average measures” type). For the total scores, we were interested in the “consistency” definition rather than absolute agreement, because total scores of the START:AV are not used in an absolute manner (i.e., as a cutoff); in fact, in clinical practice they are not used at all. For the item ratings and final risk judgments (both rated as low, moderate, or high), we calculated absolute agreement ICCs. For all ICCs values, the 95% confidence intervals (CIs) were determined. Koo and Li (2016) argue that it is more appropriate to interpret the CIs rather than the ICC point estimates. CIs give an indication of the range in which the true ICC value lands, whereas the ICC point estimate is only an expected value of the true ICC. We interpreted the CIs of the Gwet's AC1 and the single‐rater two‐way random ICCs according to Koo and Li’s (2016) guidelines: ICC < 0.50, poor; 0.50–0.75, moderate; 0.75–0.90, good; > 0.90, excellent. These interpretations are stricter than the commonly used guidelines by Landis and Koch (1977), Cicchetti and Sparrow (1981), or Fleiss (1986). It has been repeatedly argued that these classic guidelines might be too lenient for applied settings, where important decisions are made based on risk assessments (Edens & Boccaccini, 2017; Levenson, 2004). For example, the threshold of 0.80 presented by Heilbrun (1992) has been put forward as a favorable, although difficult to achieve, benchmark for field reliability in settings where the stakes are high (Edens et al., 2019; Guarnera & Murrie, 2017).

RESULTS

Forty‐six of the 100 START:AVs were completed upon admission to the service, on average, after 68 days (SD = 37, range: 0–216). The remaining 54 START:AVs in the sample were follow‐up START:AVs, completed for the purpose of treatment evaluation.

Descriptive statistics

As presented in Table 1, the evaluator groups (researchers vs. clinicians) differed significantly on the average total score for strengths [t(99) = 5.50, p < 0.001, d = 0.58] and vulnerabilities [t(99) = −7.34, p < 0.001, d = −0.71]. Research evaluators assigned lower scores for strengths (M = 14.8, SD = 5.9; clinician evaluators: M = 19.2, SD = 8.8) and higher scores for vulnerabilities (M = 36.1, SD = 7.1; clinician evaluators: M = 31.4, SD = 6.3). Across the entire sample of juveniles, the researcher‐rated total scores for strengths ranged from 3.69 to 33.58, and those for vulnerabilities ranged from 21.67 to 48.45, while the clinician‐rated total scores ranged from 5.00 to 45.76 for strengths and 16.22 to 47.00 for vulnerabilities. The average absolute difference in total scores between the research and clinical evaluators was 7.5 (SD = 5.4) for strengths, with the largest absolute difference being 27 points and three cases reaching perfect agreement. With regard to vulnerabilities, the average absolute difference was 6.6 (SD = 4.5), with the largest absolute difference being 18 points and four cases reaching perfect agreement. Lastly, Table 2 presents the prevalence of adverse outcomes in the prior and recent history of the adolescents as well as the distribution of risk estimates per outcome, as assessed by both evaluator groups.
TABLE 1

Distribution of item ratings and missing ratings for strengths and vulnerabilities for both evaluator groups, presented in percentages

START:AV itemClinical evaluatorsResearch evaluators
StrengthVulnerabilityStrengthVulnerability
Low/medium/high (L/M/H)MissingL/M/HMissingL/M/HMissingL/M/HMissing
1.School and work26/51/23011/67/22032/45/1679/43/417
2.Recreation29/57/13112/67/20126/61/9424/43/285
3.Substance use49/39/12020/46/34045/25/28240/27/330
4.Rule adherence27/58/13213/43/43138/51/1109/35/560
5.Conduct24/71/504/48/47139/60/102/29/690
6.Self‐care22/62/15110/67/21246/33/61510/49/2912
7.Coping62/37/017/21/78075/24/011/13/851
8.Impulse control56/36/5313/56/31076/17/169/30/565
9.Mental state45/46/8112/54/34062/30/358/34/553
10.Emotional state47/49/132/29/68182/15/033/17/782
11.Attitudes35/52/769/47/41368/17/0152/31/5512
12.Social skills18/65/17017/63/19133/54/8514/38/426
13.Relationships adults27/51/20212/45/41223/60/1259/38/494
13b.Relationships peers43/47/6412/53/35051/41/359/30/556
14.Social support adults41/43/15121/48/29234/41/91613/22/4916
14b.Social support peers60/26/683/49/44461/17/3195/12/6221
15.Parenting26/55/15410/41/47252/31/895/29/597
16.Parental functioning34/49/15215/49/29739/38/8159/30/4120
17.Peers74/18/359/45/42483/5/1112/23/6312
18.Material resources7/79/13124/72/317/82/748/84/44
19.Community7/84/548/85/341/96/126/89/23
20.External triggers40/28/42825/37/201852/33/21316/44/337
21.Insight43/39/13511/42/44363/30/523/33/622
22.Plans30/49/16520/57/18542/33/121318/37/2916
23.Medication adherence a 11/20/7/56612/25/3/56416/16/10/54421/14/6/545
24.Treatability28/51/16515/59/22441/44/10516/29/505
Total scores b 19.2 (8.8)31.4 (6.3)14.8 (5.9)36.1 (7.1)

Medication adherence as an additional fourth category for “not applicable” ratings.

Descriptives for total scores are presented as mean (standard deviation).

TABLE 2

Percentage of adverse outcomes rated as present in prior and recent history as well as distribution of risk ratings for both evaluator groups

Adverse outcomesClinical evaluatorsResearch evaluators
Prior historyRecent historyRisk judgmentPrior historyRecent historyRisk judgment
PresentMissingPresentMissingLow/medium/high (L/M/H)MissingPresentMissingPresentMissingL/M/HMissing
1. Violence76037039/31/29174054034/30/351
2. Nonviolent offending62134132/34/34066036037/40/230
3. Substance abuse74158225/28/45271054028/26/460
4. Unauthorized absences8826139/43/4718907108/40/520
5. Suicide24111184/11/412207083/13/40
6. NS self‐injury39128166/18/14242034068/18/140
7. Victimization82130130/35/34180045024/26/500
8. Health neglect67663726/52/18471077021/57/211

Note: NS, non‐suicidal.

Distribution of item ratings and missing ratings for strengths and vulnerabilities for both evaluator groups, presented in percentages Medication adherence as an additional fourth category for “not applicable” ratings. Descriptives for total scores are presented as mean (standard deviation). Percentage of adverse outcomes rated as present in prior and recent history as well as distribution of risk ratings for both evaluator groups Note: NS, non‐suicidal. Interrater reliability of Short‐Term Assessment of Risk and Treatability: Adolescent (START:AV) items and total scores for strengths and vulnerabilities, including interpretation of confidence intervals (CIs) Note: ICC, intraclass coefficient. Only assessed for low, moderate or high medication adherence, excluding the “not applicable” category.

Missing ratings

Table 1 also presents the percentage of missing ratings per item for both evaluator groups. Because the research evaluators relied solely on file information, they had significantly more missing items than the clinical evaluators, both for strengths [t(99) = −5.77, p < 0.001, d = −0.76] and for vulnerabilities [t(99) = −7.13, p < 0.001, d = −0.96]. The clinical evaluators had a high percentage of missing ratings on item 20, “external triggers,” whereas the research evaluators had multiple items with a relatively large number of missing ratings, especially items related to the adolescent's social context. For example, there were more than 10% missing ratings on both strengths and vulnerabilities for items 14a and 14b (social support from adults and peers), item 16 (parental functioning), and item 17 (peers). In addition, items such as self‐care (item 6), attitudes (item 11), external triggers (item 20) and plans (item 22) were among the items with the greatest number of missing ratings. With regard to the adverse outcomes, the clinical evaluators had especially high missing ratings for “health neglect” (see Table 2).

Intraclass correlation coefficients

Two‐way random absolute ICCs for the items ranged from −0.01 (external triggers) to 0.67 (parental functioning) for the strength items, and from 0.19 (peers; treatability) to 0.70 (parental functioning) for the vulnerability items (Table 3). As discussed in the Method section, we relied on the 95% CIs for interpretation because these present a more accurate picture of the true reliability scores. Almost half of the strength items (n = 12) fell within the poor reliability range and the other half (n = 13) fell within the poor to moderate range. Only for the item “parental functioning” did the evaluators reach poor to good reliability. The item‐level mean ICC for strengths was 0.31 (range: −0.01–0.67). For the vulnerability items, a similar pattern was observed: 11 items demonstrated poor reliability, 14 items poor to moderate, and again reliability of parental functioning ranged from poor to good. The item‐level mean ICC for vulnerabilities was 0.38 (range: 0.19–0.70). For the total scores, the two‐way random consistency ICC was 0.42 [95% CI 0.24–0.57] for strengths and 0.54 [0.38–0.66] for vulnerabilities, both demonstrating poor to moderate reliability (Table 3).
TABLE 3

Interrater reliability of Short‐Term Assessment of Risk and Treatability: Adolescent (START:AV) items and total scores for strengths and vulnerabilities, including interpretation of confidence intervals (CIs)

START:AV featureStrengthVulnerability
ICC95% CICI interpretation n ICC95% CICI interpretation n
1.School and work0.45[0.28–0.60]Poor to moderate930.32[0.13–0.49]Poor93
2.Recreation0.19[−0.02–0.38]Poor950.31[0.11–0.48]Poor94
3.Substance use0.38[0.20–0.53]Poor to moderate980.53[0.37–0.66]Poor to moderate100
4.Rule adherence0.44[0.27–0.59]Poor to moderate980.37[0.19–0.53]Poor to moderate99
5.Conduct0.14[−0.05–0.32]Poor1000.33[0.14–0.49]Poor99
6.Self‐care0.29[0.08–0.47]Poor840.44[0.25–0.59]Poor to moderate87
7.Coping0.33[0.14–0.49]Poor980.43[0.26–0.58]Poor to moderate99
8.Impulse control0.08[−0.10–0.27]Poor920.44[0.24–0.60]Poor to moderate95
9.Mental/Cognitive state0.42[0.22–0.58]Poor to moderate940.41[0.22–0.56]Poor to moderate97
10.Emotional state0.19[−0.01–0.38]Poor940.30[0.11–0.47]Poor97
11.Attitudes0.12[−0.06–0.30]Poor800.28[0.08–0.46]Poor85
12.Social skills0.37[0.18–0.53]Poor to moderate950.44[0.24–0.60]Poor to moderate93
13Relationships adults0.46[0.29–0.61]Poor to moderate930.47[0.30–0.62]Poor to moderate94
13b.Relationships peers0.42[0.24–0.58]Poor to moderate910.28[0.09–0.45]Poor94
14.Social support adults0.34[0.14–0.52]Poor to moderate830.34[0.14–0.52]Poor to moderate82
14b.Social support peers0.44[0.24–0.60]Poor to moderate740.31[0.09–0.50]Poor to moderate75
15.Parenting0.39[0.16–0.57]Poor to moderate870.52[0.35–0.66]Poor to moderate91
16.Parental functioning0.67[0.52–0.78]Poor–good830.70[0.55–0.81]Poor to good76
17.Peers0.46[0.27–0.62]Poor to moderate840.19[0.00–0.38]Poor84
18.Material resources0.12[−0.08–0.31]Poor960.29[0.10–0.46]Poor96
19.Community0.00[−0.20–0.20]Poor950.30[0.11–0.48]Poor94
20.External triggers−0.01[−0.26–0.25]Poor620.27[0.05–0.46]Poor76
21.Insight0.41[0.22–0.57]Poor to moderate940.33[0.14–0.50]Poor to moderate95
22.Plans0.30[0.09–0.48]Poor8350[0.31–0.65]Poor to moderate79
23.Medication adherence a 0.23[−0.13–0.53]Poor to moderate320.45[0.13–0.68]Poor to moderate31
24.Treatability0.31[0.12–0.48]Poor900.19[0.00–0.37]Poor91
Total score0.42[0.24–0.57]Poor to moderate1000.54[0.38–0.66]Poor to moderate100

Note: ICC, intraclass coefficient.

Only assessed for low, moderate or high medication adherence, excluding the “not applicable” category.

Interrater reliability of history ratings and risk judgments of the adverse outcomes with an interpretation of confidence intervals (CIs) Note: AC, agreement coefficient. As shown in Table 4, the two‐way random absolute ICCs for the final risk judgments ranged from 0.57 (health neglect) to 0.84 (suicide). For half of the risk judgments, including nonviolent offenses, substance abuse, self‐injury, and victimization, the evaluators reached moderate to good agreement. For unauthorized absences and health neglect, poor to moderate interrater reliability was found, for violence moderate reliability, and for suicide good reliability. We found category errors for all adverse outcomes, except suicide. A category error occurs when one evaluator rates the case as high risk, whereas the other evaluator rates the same case as low risk. More specifically, there were six category errors for violence, three for nonviolent offending, three for substance abuse, three for unauthorized absences, one for self‐injury, four for victimization, and one for health neglect. In half (10) of the major disagreements, the clinical evaluators assigned low risk, whereas the research evaluators estimated high risk; in the other half (11), it was the opposite. Exploratory analysis revealed no significant differences between evaluators in the prevalence of category errors.
TABLE 4

Interrater reliability of history ratings and risk judgments of the adverse outcomes with an interpretation of confidence intervals (CIs)

Adverse outcomePrior historyRecent historyRisk judgment
Gwet's AC1 (95% CI)CI interpretation n Gwet's AC1 (95% CI)CI interpretation n ICC (95% CI)CI interpretation n
1. Violence0.58 [0.42–0.75]Poor–good1000.50 [0.33–0.68]Poor to moderate1000.64 [0.51–0.74]Moderate98
2. Nonviolent offending0.57 [0.40–0.74]Poor to moderate990.68 [0.53–0.83]Moderate to good990.66 [0.52–0.76]Moderate to good100
3. Substance abuse0.77 [0.64–0.89]Moderate to good990.64 [0.48–0.67]Poor to moderate980.75 [0.65–0.83]Moderate to good98
4. Unauthorized absences0.88 [0.79–0.96]Good to Excellent980.72 [0.59–0.86]Moderate to good970.39 [0.21–0.55]Poor to moderate99
5. Suicide0.89 [0.81–0.97]Good to Excellent990.90 [0.83–0.97]Good to Excellent990.84 [0.76–0.89]Good99
6. Nonsuicidal self‐injury0.78 [0.66–0.91]Moderate to excellent990.72 [0.58–0.86]Moderate to good990.79 [0.70–0.86]Moderate to good98
7. Victimization0.72 [0.59–0.86]Moderate to good990.37 [0.18–0.56]Poor to moderate990.65 [0.50–0.76]Moderate to good99
8. Health neglect0.55 [0.38–0.73]Poor to moderate940.67 [0.51–0.82]Moderate to good930.57 [0.41–0.69]Poor to moderate95

Note: AC, agreement coefficient.

Gwet's AC1

Gwet's AC1 values were calculated for prior and recent history of adverse outcomes (see Table 4). Values ranged from 0.55 (health neglect) to 0.89 (suicide) for prior history, with wide CIs covering all levels of interrater reliability, from poor to excellent. For recent history, Gwet's AC1 ranged from 0.37 (victimization) to 0.90 (suicide), again with wide CIs. The level of interrater reliability was good to excellent for suicide, and moderate to good for most other adverse outcomes (nonviolent offending, unauthorized absences, self‐injury, and health neglect). Recent history of violence, substance abuse, and victimization demonstrated poor to moderate reliability. We also explored the number of disagreements between the evaluators for both prior (i.e., more than 4 months prior to the risk assessment) and recent history ratings, that is, how often one evaluator identified an adverse outcome in the past, while the other evaluator did not. For prior history, this occurred equally frequently. For example, in 14 cases, clinical evaluators identified violent incidents in prior history, whereas the researchers did not. Likewise, in 12 cases, research evaluators identified violent incidents in prior history, whereas the clinicians did not. For recent history, the research evaluators identified more incidents than clinical evaluators, especially for the recent history of violence (i.e., 21 identified by researchers but not identified by clinicians vs. four identified by clinicians but not identified by researchers), unauthorized absences (15 vs. 5), victimization (24 vs. 9), and health neglect (15 vs. 4).

DISCUSSION

To our knowledge, this study was the first to assess the interrater reliability of the START:AV completed by professionals in a secure youth forensic mental health setting as part of their day‐to‐day responsibilities. Consistent with emerging research on field reliability, we found lower interrater reliability than studies that examined the START:AV for research purposes only (Klimukienė et al., 2018; Viljoen et al., 2012). We first discuss our findings in light of previous research, and then describe factors that could have contributed to these lower interrater reliability findings.

Field interrater reliability of the START:AV

START:AV total scores

In line with the literature, we assessed the interrater reliability of the total scores for vulnerabilities and strengths. However, by doing so we treated an SPJ instrument actuarially, which is inconsistent with its intended use in practice (Neal et al., 2019). Therefore, our findings related to SPJ total scores are not generalizable or relevant to typical clinical practice. Nevertheless, we have included the total scores in our reliability analyses to allow for comparison with previous studies. The ICCs for both total scores were weak, with the ICC for the vulnerability total score falling barely within the moderate range and the strength total score demonstrating poor interrater reliability, according to the more stringent interpretation guidelines of Koo and Li (2016). By comparison, the total score ICCs found in the START:AV non‐field studies (Klimukienė et al., 2018; Viljoen et al., 2012) fell within the good and excellent range according to the Koo and Li guidelines. The findings in the present study are more comparable to those reported in a field study (Troquete et al., 2015) of the adult version of the risk assessment instrument, the START (Webster et al., 2009). Like us, Troquete and colleagues found a moderate interrater reliability for the vulnerability total score (ICC = 0.64) and poor interrater reliability for the strength total score (ICC = 0.49) . Other studies also corroborate our finding that reliability is lower for strengths than for vulnerabilities. For example, in a Norwegian field study with the adult START, Nonstad et al. (2010) reported that the interrater reliabilities for risks were higher than for strengths. The authors speculated that practitioners in secure settings might be more familiar with the vulnerabilities of patients than with their strengths.

START:AV items

When considering the mean item ICCs, the trend described above was also evident for the individual strength and vulnerability items in our study, despite some strengths reaching higher ICCs than their vulnerability counterparts. This is in line with Klimukienė et al. (2018), who found lower interrater reliability for strengths than for vulnerabilities for the majority of the items. Nevertheless, the ICCs for the items in the present study were considerably lower than reported in the latter study.

START:AV adverse outcomes

Central to SPJ instruments are the final risk judgments – that is, judgments about whether a person is low, moderate or high risk with respect to violence to others, nonviolent offending, substance abuse, unauthorized absences, suicide, self‐injury, victimization, and health neglect. These decisions are based on empirically derived risk factors and are to be distinguished from the total scores on strengths and vulnerabilities. Overall, evaluators in our study reached better agreement for the final risk judgments than for total scores. Also in contrast to other studies examining the START:AV (Klimukienė et al., 2018; Viljoen et al., 2012), we found that the majority of the adverse outcomes demonstrated similar or higher interrater reliability. Only the reliability of the risk of unauthorized absences was considerably lower than reported by Klimukienė et al. (2018). Of all risk estimates in their study, reliability was highest for this adverse outcome, whereas in our study, it was the lowest of all risk estimates. Specifically, the two evaluator groups in our study disagreed about whether the risk of this outcome was moderate versus high (in 28% of the cases). Further, this disagreement was not always in the same direction. In 15 cases, the research evaluator assigned high risk, while in the other 13 cases the clinical evaluator assigned high risk. We could not find a plausible explanation for this finding, except perhaps that the scoring instructions for this adverse outcome do not differentiate sufficiently between moderate and high risk. Likewise, the interrater reliability of the history ratings – which ranged from poor to excellent, with the majority of history ratings falling within the moderate range – were lower than what is typically found for static factors (Edens et al., 2019) and for historical factors in SPJ instruments in particular. For example, field studies with the Historical Clinical Risk Management‐20 (HCR‐20; Webster et al., 1997) have found good interrater reliability for the historical scale, ranging from 0.78 to 0.86 (de Vogel & de Ruiter, 2004; Jeandarme et al., 2017). However, de Vogel and de Ruiter (2004) also reported low interrater reliability for the item “previous violence.” They found that this item was misunderstood by some clinical evaluators (i.e., to not include the index offense, when it fact it does), resulting in disagreements between clinical and research evaluators. When we further explored the disagreements on history ratings, we found that research evaluators identified considerably more incidents in the recent history of several different adverse outcomes than did clinical evaluators. One hypothesis is that research evaluators were more thorough in their rating of incidents. However, we did not observe a similar pattern for prior history. An alternative explanation is that research evaluators identified more incidents because they read the daily reports and immediately documented their findings in the START:AV form, whereas clinical evaluators, at the time of the risk assessment, tended to rely more on their memory of incidents. Personal communication with treatment coordinators confirmed this hypothesis: they assigned ratings based on their recollection from observations and conversations with the youth and the team because they lacked the time to sift through the progress notes. Also noteworthy is the very low interrater reliability on the final risk judgment of health neglect, an issue not addressed in any other study. The ICC for this scale was one of the lowest we found. Health neglect behaviors may be more challenging to rate than, for example, violent or suicidal behaviors. Evaluators may have differing understandings of the extent to which behaviors such as insufficient sleep, unhealthy diet, or smoking should be considered as health neglect. In addition, discrepancies in the information available to the evaluator groups may have contributed to a poor interrater reliability. Research evaluators had more missing ratings on the item involving self‐care, which directly informs the adverse outcome of health neglect. Information required to rate self‐care and the associated risk was likely not sufficiently documented in the progress notes, whereas clinicians could rely on their own or their team's observations.

Factors influencing interrater reliability in the field

Confounds are an integral part of field studies and represent the realities of conducting risk assessment in practice settings. The goal is not to eliminate them, but to explore their influence on reliability (Guarnera & Murrie, 2017). We discuss factors that may have affected the present field study, resulting in lower interrater reliability than reported in the START:AV non‐field studies, especially for the items and total scores.

Factors related to the study design

The START:AV non‐field studies (Klimukienė et al., 2018; Viljoen et al., 2012) likely benefited from a research design that required evaluators to reach sufficient reliability prior to starting the actual data collection. Such a procedure promotes consistency across the evaluators from the outset. In the present study, we did not establish interrater reliability for either the clinical evaluators or the research evaluators prior to data collection. This likely reflects most real‐world services, which have to deal with time and resource constraints. Nevertheless, service providers and agencies are advised to conduct their own reliability and validity check of the risk assessment instrument prior to full implementation and to continue to assess its performance periodically (Vincent et al., 2012). The present study also differed from the START:AV non‐field studies with respect to the time between the training and the data collection. In our study, the last training moment (i.e., feedback on a practice case) occurred 3 months prior to the first START:AV assessments. In fact, 17 months had passed since the initial workshop on the administration of the START:AV. Clinicians have discussed this gap between training and the actual use of the START:AV as a barrier to the risk assessment implementation (De Beuf et al., 2020). In most non‐field studies, this gap is smaller, because data collection usually starts shortly after training. This is relevant because over time, more errors may occur due to rater drift, hence weakening reliability (Vincent et al., 2012). In an early field study with the HCR‐20, Belfrage (1998) actually excluded an evaluator from the reliability analyses, because this person scored significantly differently from the other clinicians, arguably due to a gap of 6 months between the training and the risk assessments. That said, in the present setting, some evaluators participated in a refresher workshop that took place halfway through the data collection period (De Beuf et al., 2020). This may have reduced rater drift, although the sample was too small to explore this further. Another notable real‐world factor that was at play in our field study was the variance in information sources between the evaluators. To complete the risk assessments, both evaluator groups had access to the same file information. However, the clinical evaluators additionally had direct contact with the adolescent and their treatment team. This may have resulted in qualitatively and quantitatively different input from both evaluators (Penney et al., 2014). For example, in the study conducted by de Vogel and de Ruiter (2004), treatment supervisors reported that they did not feel the need to review file information as they felt that this information was already known to them. They stated that they mainly relied on personal interactions with the patient. The treatment coordinators in the present setting confirmed this: for the sake of efficiency, they relied on their own observations and personal interactions. This approach to risk assessment by clinical staff may also explain why we found lower reliability for history ratings. Clinicians can have various reasons for not relying on file information, such as time constraints, negligence, overestimation of one's knowledge of the file, or simply because they did not adhere to the user manual (Storey et al., 2012).

Factors related to the instrument

Rufino et al. (2011) found inverse correlations between interrater agreement and subjectivity of PCL‐R and HCR‐20 items. Similarly, other authors have found lower reliability for items that require more clinical judgment (Kennealy et al., 2016; Quesada et al., 2014). This conclusion might also be applicable to the START:AV; except for the history of adverse outcomes, the instrument consists of dynamic items that require clinical judgment. Ratings of low, moderate, or high are assigned based on the presence of “item anchors” (descriptors that represent the item) such as “acts without thinking” (item 7: impulsivity) or “caregivers provide a structured and stable environment” (item 15: parenting). However, these item anchors do not have explicit cutoff points to guide an evaluator in differentiating between, for example, a moderate and a high rating. The evaluator has discretion to rate the items as low, moderate, or high on the basis of the presence and severity of the item anchors observed in the assessed adolescent. The absence of clear boundaries between the ratings creates an opportunity for subjectivity. Thus, establishing behaviorally anchored cutoffs may help reduce subjectivity. Dynamic items such as those found in the START:AV might be challenging to complete based on file information only (Penney et al., 2014). Indeed, the research evaluators in our study had considerably more missing item ratings than the clinical evaluators. Items that involved insight into the adolescent's situation outside the treatment setting (i.e., social support, peers, parental functioning, external triggers) had between 13% and 21% missing ratings. Similarly, items such as attitudes, plans, and self‐care had considerably more missing ratings when completed by researchers, probably because this information was less often included in the daily progress notes.

Practical implications

Aiming for consensus

Our study indicates that, although they share some agreement, the two evaluator groups held different pieces of the risk assessment puzzle. Research evaluators extracted more incidents that occurred in the facility (i.e., recent history) from daily reports, whereas clinical evaluators relied on their interactions with the adolescent and team when rating the dynamic items. The evaluator groups not only brought different information to the table, they also differed in their relationship with the adolescent. The research evaluators had a more distant, objective position, whereas the clinicians, although serving in a supervisory and decision‐making role rather than a therapeutic role, were more closely involved with the youth. Clinicians' feelings, positive and negative, that come with this relationship may result in subjectivity and impact the risk assessment (de Vogel & de Ruiter, 2004). Nevertheless, de Vogel and de Ruiter (2006) demonstrated that risk assessments of both evaluator groups did not differ in terms of predictive performance. To optimally benefit from both information sources, the consensus approach is recommended (de Vogel & de Ruiter, 2004). In this approach, multiple staff members (e.g., researcher, treatment coordinator) complete the START:AV independently of each other and discuss their findings during a meeting, resulting in a consensus START:AV. Although the effect of the consensus approach on interrater reliability has, to our knowledge, not yet been evaluated, a study with the HCR‐20 has demonstrated that a consensus assessment produced the highest predictive accuracy for future recidivism (de Vogel & de Ruiter, 2006). Agencies and services that prepare to introduce SPJ risk assessment instruments into their routine practice should consider integrating the consensus method in the implementation plan. Settings that do not have researchers or other objective evaluators available could create a rotation system of fellow clinicians to serve as second independent raters who complete risk assessments based on daily reports and other file information.

Other suggestions for practice

In the event that consensus meetings may be challenging to organize due to limited resources or other constraints, service providers could consider alternatives, such as regular case discussions and peer review. A one‐time training in the use of the risk assessment instrument is not sufficient to secure reliability because rater drift may occur. Vincent et al., (2012) suggested periodical monitoring to identify challenges in reliability and maintain a satisfactory level of rating consistency. Reliability may be considered a result of implementation fidelity, which is the degree to which a practice is being delivered as intended by its developers (Proctor et al., 2011). In the intervention literature, researchers have developed checklists to assess implementation fidelity of a practice of interest. Likewise, we developed a scale to assess the level of completion of START:AV rating forms: the START:AV Adherence Rating Scale (STARS; De Beuf, de Vogel, & de Ruiter, 2020). This scale codes how complete the rating forms are, but does not investigate the quality of the ratings as measured by the descriptors listed in the user guide. While the latter measure would obviously be most relevant to the concept of reliability, the extent to which the adherence scale could serve as a proxy for reliability should be explored. If the STARS score indeed correlates with reliability, it would present a quick and accessible method to periodically assess this psychometric quality. In addition to training and monitoring efforts, well‐communicated arrangements about which information is to be used for the risk assessment, taking feasibility into account, may benefit the reliability of the ratings. Moreover, service providers that are unfamiliar with documenting strengths may want to invest additional effort in coaching staff on the observation and documentation of strengths and talents of clients. This advice also applies to items that are routinely found to be absent in files (e.g., attitudes and self‐care in the present setting). Risk assessment tool developers and translators could help to improve reliability by presenting concrete examples and unambiguous scoring instructions in the user guide. As mentioned earlier, several START:AV rating instructions allow room for subjective opinion, which could be limited by adding more examples (e.g., examples of mild health neglect, yet serious enough to rate as present in history). It is recommended that developers of risk assessment instruments continue to collect feedback from practitioners and trainees, because it may provide helpful suggestions for fine‐tuning of the instrument's instructions.

Limitations and future directions

Field interrater reliability ideally involves two field evaluators who are both familiar with the examinee and who conduct the assessment for clinical‐forensic decision‐making purposes, based on the same information. However, in clinical practice, it is not that common to have two or more independent evaluators conduct the same risk assessment. Hence, in our study, we relied on research assistants as the second evaluator of the START:AV. This limits the interpretation of the study as a prototypical field study, where both assessments are completed for non‐research purposes. Our study is also limited in that the evaluator groups had different information at their disposal. The clinical evaluators could rely on direct and indirect observations of the adolescent, in addition to file information, while the researchers had access only to the latter source. Nevertheless, even in practice, evaluators rarely will collect and use the exact same information. Therefore, the present study provided an opportunity to identify strengths in the approach of each evaluator group. Further, as discussed earlier, the consensus methodology is a way to turn these information differences into a strength by bringing together multiple perspectives in consensus meetings. Moreover, it would be informative to investigate in which direction the consensus ratings are trending across and within evaluators, compared with the individual ratings. Multiple evaluator characteristics (e.g., training, profession, experience, personality traits) can affect interrater reliability (Boccaccini et al., 2008). Not assessing these variables can be considered a limitation of this study. Future research should aim to measure these and other possibly confounding variables to be better able to explain whether the variance in interrater reliability is true variance or error due to evaluator and context effects. By increasing the number of available field studies and aggregating their findings, we may eventually be more confident in gauging the performance of the risk assessment instruments in real‐world settings. Furthermore, future (field) research should invest in assessing whether evaluators identify the same items or item clusters as key or critical. Ultimately, these strengths and vulnerabilities are most likely to inform risk management goals. If the determination of key and critical items demonstrates low reliability, their relevance to risk management may be questionable. Likewise, the reliability of risk formulations, scenario planning, and—most critically—risk management plans is a relevant topic for future research. Another important point for future research is that researchers specify which interrater reliability statistic was used, because this affects the interpretation and comparisons that can be made. For example, there are 10 forms of ICCs and the reader should be able to verify whether the appropriate ICC was used. Likewise, it is recommended to report the CIs of reliability statistics because they present a more complete picture of the estimate, resulting in more appropriate conclusions about the level of reliability (Koo & Li, 2016). In the present study, the CIs are relatively wide, which is in line with Edens and Boccaccini's (2017) comment that with lower reliability come wider CIs. In four instances, the CI covered three qualitative labels (e.g., the reliability of prior history of Violence ranged from poor to good). Edens and Boccaccini (2017) anticipated that CIs would be wider in field studies than non‐field studies. However, we could not confirm this for the START:AV, as the previous studies have not reported CIs for interrater reliability. The field should also continue the conversation on what is an acceptable level of reliability in legal and forensic settings. Currently, there is no agreement on a benchmark for sufficient reliability of forensic risk assessment instruments in the field, nor for other forensic evaluations (Edens & Kelley, 2017). Even in forensic sciences, there is currently no standard of acceptable error rates (Dror, 2020). Establishing what is “acceptable” is a complex matter, because it is specific to a forensic domain, setting, evaluator, case, and so on (Dror, 2020). Conducting more field studies might give direction as to what is reasonable to expect, what is possible to achieve, and what is minimally required considering the high stakes of risk assessment decisions in the forensic and legal arena.

CONCLUSIONS

Compared with the sheer volume of studies on the validity of risk assessment instruments, reliability studies are underrepresented in the peer‐reviewed literature. This is especially true for field studies on reliability. The present field study addressed this gap by evaluating the START:AV in a secure youth care facility. We found moderate to good interrater reliability for most final risk judgments, which, based on how they are intended to be used clinically, are the most essential decisions made during an SPJ risk assessment before proceeding to risk management. Still, the interrater reliability for the individual items and total scores was lower than previously found in START:AV non‐field studies. This is in line with the growing literature on risk assessment instruments in real‐world settings. Understanding an instrument's performance in a particular setting is relevant to agencies and services that wish to implement it. Implementation plans should take into account real‐life factors that may affect interrater reliability and take steps to minimize their influence. For example, the consensus approach to risk assessment may strengthen field reliability, which in turn may enhance field validity. That is, the “boss” showing who's boss.
  26 in total

1.  Field reliability of the SAVRY with juvenile probation officers: Implications for training.

Authors:  Gina M Vincent; Laura S Guy; Samantha L Fusco; Bernice G Gershenson
Journal:  Law Hum Behav       Date:  2012-06

2.  Sexual predator civil commitment: a comparison of selected and released offenders.

Authors:  Jill S Levenson
Journal:  Int J Offender Ther Comp Criminol       Date:  2004-12

3.  The Short-Term Assessment of Risk and Treatability (START): a prospective validation study in a forensic psychiatric sample.

Authors:  Tonia L Nicholls; Johann Brink; Sarah L Desmarais; Christopher D Webster; Mary-Lou Martin
Journal:  Assessment       Date:  2006-09

4.  An examination of the interrater reliability between practitioners and researchers on the static-99.

Authors:  Stephen P Quesada; Cynthia Calkins; Elizabeth L Jeglic
Journal:  Int J Offender Ther Comp Criminol       Date:  2013-07-16

5.  Taking forensic mental health assessment "out of the lab" and into "the real world": Introduction to the special issue on the field utility of forensic assessment instruments and procedures.

Authors:  John F Edens; Marcus T Boccaccini
Journal:  Psychol Assess       Date:  2017-06

6.  Actuarial risk assessment of sexual offenders: The psychometric properties of the Sex Offender Risk Appraisal Guide (SORAG).

Authors:  Martin Rettenberger; Marnie E Rice; Grant T Harris; Reinhard Eher
Journal:  Psychol Assess       Date:  2017-06

7.  The Error in "Error Rate": Why Error Rates Are So Needed, Yet So Elusive.

Authors:  Itiel E Dror
Journal:  J Forensic Sci       Date:  2020-04-21       Impact factor: 1.832

8.  Developing criteria for establishing interrater reliability of specific items: applications to assessment of adaptive behavior.

Authors:  D V Cicchetti; S A Sparrow
Journal:  Am J Ment Defic       Date:  1981-09

9.  Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda.

Authors:  Enola Proctor; Hiie Silmere; Ramesh Raghavan; Peter Hovmand; Greg Aarons; Alicia Bunger; Richard Griffey; Melissa Hensley
Journal:  Adm Policy Ment Health       Date:  2011-03

10.  Pilot Implementation and Preliminary Evaluation of START:AV Assessments in Secure Juvenile Correctional Facilities.

Authors:  Sarah L Desmarais; Brian G Sellers; Jodi L Viljoen; Keith R Cruise; Tonia L Nicholls; Joel A Dvoskin
Journal:  Int J Forensic Ment Health       Date:  2012-11-06
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.