Literature DB >> 35494923

The Impact of Electronic Data to Capture Qualitative Comments in a Competency-Based Assessment System.

Teresa M Chan1,2, Stefanie S Sebok-Syer3, Yusuf Yilmaz2,4, Sandra Monteiro5.   

Abstract

Introduction Digitalizing workplace-based assessments (WBA) holds the potential for facilitating feedback and performance review, wherein we can easily record, store, and analyze data in real time. When digitizing assessment systems, however, it is unclear what is gained and lost in the message as a result of the change in medium. This study evaluates the quality of comments generated in paper vs. electronic media and the influence of an assessor's seniority. Methods Using a realist evaluation framework, a retrospective database review was conducted with paper-based and electronic medium comments. A sample of assessments was examined to determine any influence of the medium on the word count and the Quality of Assessment for Learning (QuAL) score. A correlation analysis evaluated the relationship between word count and QuAL score. Separate univariate analyses of variance (ANOVAs) were used to examine the influence of the assessor's seniority and medium on word count, QuAL score, and WBA scores. Results The analysis included a total of 1,825 records. The average word count for the electronic comments (M=16) was significantly higher than the paper version (M=12; p=0.01). Longer comments positively correlated with QuAL score (r=0.2). Paper-based comments received lower QuAL scores (0.41) compared to electronic (0.51; p<0.01). Years in practice was negatively correlated with QuAL score (r=-0.08; p<0.001) as was word count (r=-0.2; p<0.001). Conclusion Digitization of WBAs increased the length of comments and did not appear to jeopardize the quality of WBAs; these results indicate higher-quality assessment data. True digital transformation may be possible by harnessing trainee data repositories and repurposing them to analyze for faculty-relevant metrics.
Copyright © 2022, Chan et al.

Entities:  

Keywords:  digitizing; electronic comment; paper-based comment; realist evaluation framework; workplace-based assessment

Year:  2022        PMID: 35494923      PMCID: PMC9038604          DOI: 10.7759/cureus.23480

Source DB:  PubMed          Journal:  Cureus        ISSN: 2168-8184


Introduction

Electronic data capture systems, which are the new normal at most of the educational institutions that collect assessment data about trainees [1,2], require careful transition planning and change management planning [3,4]. While the previous studies provide some good examples of transitions from a technical standpoint, there may be important individual differences in how faculty adapt to these changes, which may, in turn, lead to different interpretations of educational assessment outcomes [1-4]. Electronic assessment records have the potential to drastically improve assessment data aggregation. If properly implemented, they hold the potential for facilitating feedback and performance review, wherein we can easily record, store, and analyze data in real time [1,5]. Sentiment analysis or natural language machine learning algorithms hold great promise in helping to enhance real-time qualitative analysis as well [6,7], but these technologies are contingent on raters submitting high-quality observations and assessments. Specifically, faculty may differ in how they interact with electronic assessment platforms when recording feedback to trainees [8]. Critically, the technology used to gather and house assessment data may influence the rater-trainee experience. It is well described that electronic medical records (EMRs), for instance, have greatly changed the physician-patient relationship [9]. Similarly, workplace-based assessment (WBA) systems will likely affect the bedside learning environment for both teachers and trainees. For example, paper-based assessments may enhance the timeliness of feedback at the bedside (e.g., Entrustable Professional Activities, Daily Encounter Cards) but are notoriously cumbersome to aggregate [10]. Certainly, communication scientists have established that the change in a medium can greatly affect the message contained within, both in the way that people communicate these messages and the way that they are received [11]. Thus, the bigger challenge is ensuring robust data collection that is not subject to external influences, such as an awkward data entry system or technological barriers. While these considerations are important as part of routine quality assurance processes, strong anecdotal evidence highlighted a need to investigate the contextual influence of the medium on the quality of the feedback messages. The McMaster Modular Assessment Program (McMAP) was created as a workplace-based assessment program by the emergency medicine training program as a pilot competency-based medical education program [12]. With three levels (junior, intermediate, and senior) of progressively difficult task-based assessments, this program scaffolded tailored learning experiences and competency-focused instruction for trainees in a planned progression within a program of assessment [12]. After a year of piloting a WBA program of assessment with paper booklets in 2012-2013 [12], our program transitioned away from paper to an electronic version of data collection for logging our WBAs in July 2013 with customized branching online data collection forms via our institutional Medportal system (powered by Google Forms™ (Google LLC, Mountain View, California, United States)). Each trainee had their own suite of personalized Google Forms to collect workplace-based assessments, which were then managed manually by our lead designer (TC). After the transition to the new electronic system, in a subsequent quality improvement focus group, our local residents postulated that the implementation of an electronic system changed the nature of the feedback they received. Specific concerns were raised that more senior faculty provided fewer detailed comments within an electronic platform compared to the previous paper-based platform [8]. At first glance, this concern contradicts recent work. Govaerts and colleagues have noted the effects of a rater’s expertise on assessments; namely that with more complex behaviors, experienced raters tended to take longer to consider the information, searching for additional cues and observing trainees for longer [13]. Experienced raters tended to provide more interpretative, inferential judgments, whereas novice raters tended to provide more literal descriptions. Also, expert raters were thought to have superior abilities to analyze and evaluate contextual or situation-related cues [13]. However, it is unclear whether the more thoughtful analytic approach of expert raters is reflected in word count and comment quality. The main purpose of the current study was to evaluate if the quality of comments generated in paper vs. electronic media was influenced by an assessor’s seniority. Specifically, we sought to examine the effect of seniority on the quantity (as measured by word count) and quality of written feedback, as assessed using the Quality of Assessment for Learning (QuAL) rubric [14]. We evaluated whether this influence differed for paper and electronic assessments.

Materials and methods

We adopted a realist evaluation perspective to examine our local workplace-based assessment program [15]. The goal of a realist evaluation framework is to take into consideration how a program’s implementation is affected by contextual factors and how that context works in conjunction with a given mechanism to result in outcomes. Locally, at McMaster University’s specialist emergency medicine program, we have developed a daily WBA program known as the McMaster Modular Assessment Program (McMAP) [12]. Details regarding our successful blueprinting and implementation are discussed elsewhere [12,16-18]. The current study contributes to continued quality assurance and explores the contextual influence of converting McMAP to computer-based data collection methods. We are pleased to report that the previous version of this work was presented as a virtual poster at the Association for Medical Education in Europe (AMEE) 2020 virtual conference. Procedure We retrospectively examined the assessments from 85 raters on 30 residents from October 2012 to June 2015. The data from these assessments were preprocessed to anonymize the comments, names, and any user identifier and securely stored in our lead author's encrypted university computer. Comments were then evaluated to objectively determine a quality score and word count. Data selection The comments from October 2012 to June 2013 were collected on a paper-based series of WBA workbooks, which were filled out contemporaneously in the clinical setting. The comments from July 2013 to June 2015 were collected using an online e-portfolio system created and housed locally at McMaster University. All data were entered into a Microsoft Excel workbook (Microsoft Corp., Seattle, WA). Each comment examined was associated with a word count, which was determined using standard data processing functions within Excel. To calculate the number of words using Excel, we first removed double spaces and counted the spaces in the comment and added one. Records with zero to three words were excluded from this study. Appendix Table 3 outlines the comments that only had zero to three words so that our readers can appraise their worth as needed.
Table 3

Study dataset

Above are the comments for each of our excluded comments (e.g., less than four words). Please note that some of the fields were blank (in paper format) and were bypassed in the electronic format (by placing a space " ") to bypass the mandatory field requirements. The "??" indicates a word that our transcribers could not discern from the handwritten comments. The following quotes are direct verbatim statements written by our raters. The following are definitions that we have inferred but not confirmed: McMAP: McMaster Modular Assessment Program, QuAL: Quality of Assessment of Learning, n/a: not applicable, H/P/P/A: History/Physical/Plan/Assessment, RI: first-year resident, PGY: postgraduate year, ER: emergency room, prev.: previous, mgmt: management, dept: department.

IDCommentWord CountMcMAP ScoreQuAL ScorePaper = 1; Electronic =0
1n/a0601
2n/a0600
3n/a0601
4 0 01
5n/a0600
6n/a0700
7n/a0600
8n/a0601
9n/a0601
10n/a0601
11n/a0601
12n/a0601
13n/a0601
14n/a0501
15n/a0601
16n/a0601
17n/a0601
18n/a0601
19n/a0601
20n/a0600
21n/a0601
22n/a0601
23n/a0601
24n/a0601
25n/a0601
26n/a0601
27n/a0601
28n/a0601
29n/a0601
30n/a0601
31n/a0600
32n/a0600
33n/a0601
34n/a0601
35n/a0601
36n/a0601
37n/a0601
38n/a0501
39n/a0601
40n/a0601
41n/a0601
42n/a0501
43n/a0501
44n/a0600
45n/a0700
46n/a0700
47n/a0700
48n/a0500
49n/a0600
50n/a0600
51n/a0600
52n/a0600
53n/a0600
54n/a0600
55n/a0600
56n/a0700
57n/a0500
58n/a0600
59n/a0500
60n/a0600
61n/a0600
62n/a0600
63n/a0600
64n/a0700
65n/a0700
66n/a0700
67n/a0500
68n/a0600
69n/a0600
70n/a0600
71n/a0600
72n/a0600
73n/a0600
74n/a0600
75n/a0600
76n/a0600
77n/a0601
78n/a0700
79n/a0601
80n/a0600
81n/a0601
82n/a0600
83n/a0600
84n/a0600
85n/a0600
86n/a0500
87n/a0600
88n/a0600
89n/a0600
90n/a0600
91n/a0600
92n/a0600
93n/a0600
94n/a0600
95n/a0600
96n/a0500
97n/a0600
98n/a0500
99n/a0500
100n/a0600
101n/a0600
102n/a0600
103n/a0600
104n/a0500
105n/a0500
106n/a0600
107n/a0500
108n/a0600
109n/a0600
110n/a0600
111n/a0401
112n/a0701
113n/a05.501
114n/a0501
115n/a0600
116n/a0501
117n/a0501
118n/a0501
119n/a0601
120n/a0601
121n/a0501
122n/a0601
123n/a0601
124n/a0701
125n/a05.501
126n/a0600
127n/a0500
128n/a0700
129n/a0600
130n/a0600
131n/a0700
132n/a0600
133n/a0600
134n/a0600
135n/a0701
136n/a0701
137n/a0601
138n/a0601
139n/a0701
140n/a0600
141n/a0400
142n/a0701
143n/a0701
144n/a0701
145n/a0601
146n/a0700
147n/a0701
148n/a0601
149n/a0701
150n/a0700
151n/a0700
152n/a0700
153n/a0400
154n/a0501
155n/a0700
156n/a0601
157n/a0700
158n/a0600
159n/a0600
160n/a0700
161n/a0700
162n/a0700
163n/a0701
164n/a0601
165n/a0700
166n/a0700
167n/a0700
168n/a0700
169n/a0701
170n/a0701
171n/a0701
172n/a0601
173n/a0700
174n/a0700
175n/a0700
176n/a0700
177n/a0700
178n/a0600
179n/a0601
180n/a0500
181n/a0501
182n/a0601
183n/a0601
184n/a0601
185n/a0501
186n/a0600
187n/a0600
188n/a0600
189n/a0501
190n/a0501
191n/a0601
192n/a0600
193n/a0601
194n/a0601
195n/a0601
196n/a0600
197n/a0600
198n/a0600
199n/a0600
200n/a0600
201n/a0700
202n/a0601
203n/a0501
204n/a0501
205n/a0501
206n/a0401
207n/a0501
208n/a0501
209n/a0600
210n/a0601
211n/a0401
212n/a0401
213n/a0500
214n/a0600
215n/a0500
216n/a0600
217n/a0500
218n/a0600
219n/a0700
220n/a0601
221n/a0701
222n/a0601
223n/a0601
224n/a0601
225n/a06.501
226n/a0600
227n/a0500
228n/a0600
229n/a0500
230n/a0600
231n/a0600
232n/a0600
233n/a0601
234n/a0 01
235n/a0601
236n/a0601
237n/a0 01
238n/a0601
239n/a0600
240n/a0501
241n/a0601
242n/a0601
243n/a0600
244n/a0601
245n/a0601
246n/a06.501
247n/a0601
248n/a0601
249n/a0601
250n/a0601
251n/a0501
252n/a0401
253n/a0601
254n/a0501
255n/a0600
256n/a0401
257n/a0501
258n/a0501
259n/a0500
260n/a0401
261n/a0501
262n/a0501
263n/a0601
264n/a0500
265n/a0500
266n/a0600
267n/a0600
268n/a0600
269n/a0701
270n/a0601
271n/a0600
272n/a0700
273n/a0501
274n/a0601
275n/a0401
276n/a0401
277n/a0501
278n/a0501
279n/a0701
280n/a0601
281n/a0501
282n/a0500
283n/a0600
284n/a0600
285n/a0600
286n/a0600
287n/a0600
288n/a0500
289n/a0500
290n/a0500
291n/a0600
292n/a0500
293n/a0400
294n/a0600
295n/a0600
296n/a0600
297n/a0500
298n/a0600
299n/a0700
300n/a0600
301n/a0600
302n/a0700
303n/a0600
304n/a0700
305n/a0600
306n/a0600
307n/a0501
308n/a0501
309n/a0501
310n/a0501
311n/a0600
312n/a0400
313n/a0500
314n/a0600
315n/a0600
316n/a0600
317n/a0700
318n/a0700
319n/a0600
320n/a0600
321n/a0601
322n/a0600
323n/a0700
324n/a0700
325n/a04.501
326n/a0501
327n/a0601
328n/a0601
329n/a0601
330n/a0501
331n/a05.501
332n/a0501
333n/a0501
334n/a0601
335n/a0601
336n/a0401
337n/a0500
338n/a0401
339n/a0501
340n/a0501
341n/a0601
342n/a0601
343n/a0401
344n/a0400
345n/a0600
346n/a0600
347n/a0600
348n/a0701
349n/a0601
350n/a0601
351n/a0601
352n/a0501
353n/a0500
354n/a0501
355n/a0601
356n/a0500
357n/a0501
358n/a0501
359n/a0501
360n/a0601
361n/a0501
362n/a0601
363n/a0600
364n/a0700
365n/a0700
366n/a0601
367n/a0600
368n/a0600
369n/a0600
370n/a0600
371n/a0600
372n/a0500
373n/a0600
374n/a0700
375n/a0500
376n/a0500
377n/a0600
378n/a0600
379n/a0600
380n/a0700
381n/a0700
382n/a0600
383n/a0600
384n/a0600
385n/a0600
386n/a0600
387n/a0600
388n/a0600
389n/a0600
390n/a0600
391n/a0600
392n/a0600
393n/a0600
394n/a0500
395n/a0500
396n/a0600
397n/a0701
398n/a0701
399n/a0500
400n/a0700
401n/a0600
402n/a0701
403n/a0501
404n/a0701
405n/a0601
406n/a0601
407n/a0701
408n/a0701
409n/a0701
410n/a0700
411n/a0701
412n/a0500
413n/a0600
414n/a0500
415n/a0400
416n/a0601
417n/a0601
418n/a0601
419n/a0701
420n/a0501
421n/a0601
422n/a0601
423n/a0701
424n/a0701
425n/a0601
426n/a0501
427n/a0501
428n/a0601
429n/a0601
430n/a0501
431n/a0601
432n/a0501
433n/a0501
434n/a0601
435n/a0601
436n/a0601
437n/a0601
438n/a0600
439n/a0601
440n/a0601
441n/a0601
442n/a0601
443n/a0701
444n/a0601
445n/a0601
446n/a0601
447n/a0701
448n/a0700
449n/a0500
450n/a0600
451n/a0500
452n/a0700
453n/a0600
454n/a0 01
455n/a0 00
456n/a0 00
457n/a0701
458n/a0601
459n/a0601
460n/a0601
461n/a0601
462n/a0401
463n/a0500
464n/a0600
465n/a0600
466n/a0600
467Excellent1601
468?Rushes1601
469None1601
470None1601
471Good!!!1700
472nil1701
473Excellent1600
474Awesome.1701
475able1600
476Great1700
477Accurate1611
478Excellent!1601
479none1701
480Superlative1700
481Exceptional1700
482Solid1601
483No concerns.2600
484Excellent H/P/P/A2501
485Progressing well2601
486Excellent work2601
487See previous2500
488No concerns2601
489no concerns2600
490no concerns2600
491no concerns2600
492as above2600
493Neonatal jaundice2611
494Document assessment2600
495see prev.2600
496See previous2600
497No weaknesses2601
498See previous.2600
499see above..2700
500see prev.2700
501No concerns2501
502No concerns.2500
503No concerns2601
504Excellent resident2701
505as above2600
506Good resident2700
507as above2700
508Strong resident!2701
509Excellent resident2701
510Simply outstanding !!!2700
511No concerns.2701
512Excellent job2701
513overall efficient2601
514overall good.2600
515No concern2701
516Excellent resident!2701
517As above2600
518See above2700
519Works independently25.501
520Works independently2511
521Works efficiently2611
522Works independently2601
523Strong RI2601
524Efficient historian.2601
525Great documentation2701
526See above2700
527Efficient, thorough.2700
528Great charting2611
529Excellent intubation2610
530Hard worker2411
531Progressing well2601
532Good resident2611
533Excellent resident2701
534Good instructions2601
535Excellent resident2701
536Hard worker2601
537Working well2600
538Excellent resident2600
539No concerns2600
540meets criteria2700
541Good shift2701
542As above.2700
543Great shift2700
544As above.2700
545Become PGY3!2701
546No concerns2501
547As above2600
548Excellent resident2600
549Good resident.2600
550see above2600
551see above2500
552progressing well2600
553good assessments2600
554meets requirements2700
555Good assessments.2500
556Functions well2600
557excellent assessments2600
558good shift2600
559strong resident2600
560Well done.2600
561competent resident2600
562Technologically sound2701
563Doing well.2700
564Excellent work2601
565Awesome work!2701
566Excellent work!2701
567Excellent job.2600
568Well done.2600
569Good resident2601
570Overall good.2600
571Great suturing!2601
572Appropriate investigations2501
573No concerns.2600
574doing well2600
575doing well2700
576doing ok2500
577as above2700
578As above2700
579as above2700
580as above2700
581as above2700
582Good charting2711
583No concerns2601
584Competent/efficient2701
585Well done2601
586Solid shift2700
587As above2400
588Strong resident.2600
589as above.2500
590as previous.2700
591As before.2700
592solid shift.2700
593Great job.2700
594No concerns.2600
595Good shift2500
596excellent shift2600
597done previously2700
598Good assessments2611
599Excellent assessment2711
600Independent, knowledgeable2601
601Functions independently2601
602Good assessments2601
603Excellent assessments2701
604Great assessments.2510
605Good job2601
606See previous2600
607As above2600
608See previous2600
609Pleasant, motivated2601
610Excellent shift2701
611Excellent shift2701
612Excellent shift2701
613Excellent shift2701
614Excellent shift2701
615Excellent shift2701
616Excellent shift2600
617Consistently professional25.801
618doing well2600
619See consultation comments3500
620Excellent work today3401
622Excellent efficient resident.3610
623My first one3700
624See previous note3600
625done as above3700
626Good shift. Busy.3601
627As previously noted3600
628Functions well independently.3500
629Very enthusiastic. Hardworking3601
630Document pertinent negatives3 10
631Good differential diagnosis3600
632Good presentation skills3510
633see prev. explanation..3600
634appropriate for level3500
635Good patient advocacy3710
636Excellent, excellent resident!3701
637See previous comments3700
638Doing very well.3600
639Excellent management plans.3600
640Great! Superior resident3701
641Functions independently, efficient3601
642good resident independent3600
643Runs department effectively3700
644Independent and confident.3710
645Efficient consent obtained3711
646No concerns. Excellent3601
647Appropriate for senior3701
648Proactive and enthusiastic3601
649Review toxic alcohols3611
650pleasure, developing nicely.3600
651Not grossly unsatisfactory.3700
652Hardworking, no concerns3601
653Very good resident3601
654Student progress well3501
655Hard working resident3701
656hard working resident,3610
657Continues? comprehensive assessment3611
658Consistently good performance3601
659Structure differential/plan3411
660Short ?? ??3601
661Organized/appropriate plans3501
662Follow-up/reassess ??3401
663Good overall performance3500
664Consistent comprehensive performance3610
665Well done. Excellent.3701
666Go to PGY33701
667Time for PGY33701
668Very strong resident3610
669Very good resident.3600
670Very good resident.3600
671Very good resident3600
672Overall very good.3600
673Good job today3501
674Great job today!3600
675Independent with procedure3700
676strong clinical skills3600
677Excellent patient care.3710
678Great work today!3601
679Efficient, independent, multitasking.3600
680Overall excellent resident3601
681Overall great resident36.501
682Overall very good3600
683continuing to improve3500
684Excellent. Dependable. Trustworthy.3700
685Very strong resident.3700
686A good day3701
687Another great shift3701
688A good shift.3600
689not completed today3500
690No real concerns3600
691Managed volume well3701
692Very good assessments3611
693Very good assessments3611
694Better organized today3601
695Read around cases3501
696functions very well.3600
697Good ER shift3701
698Excellent ?? shift3701
699Excellent mgmt/presentation3701
700Appropriate for level3401
701Excellent  overall shift3701
702Good shift today3701
703Very good shift3601
704Appropriate for PGY23500
705Excellent overnight shift...3600
706Excellent independent resident3701
708Carried dept - mostly3710
709doing very well3600
710continues to progress3600
711MOTIVATED, WORKS HARD3610
712did very well3600
713worked independently well3700
  Mode600
    Average=0.08Count paper=338
   Count electronic=373
Coding There were two independent variables: method of data capture and years in practice. Method of data capture was medium, coded as paper or electronic entry. Years in practice was coded as a continuous variable. Dependent variables were word count, coded as a count for each comment, and comment quality, coded using the QuAL score [14] and McMAP global scale [12], which is a global rating scale of trainee performance on a particular shift rated out of seven on behavioral anchors. Average word counts were rounded and reported as whole numbers [12]. As a result of excluding records with lower word counts, there were no missing data for word count or for QuAL score. However, there were seven missing McMAP global scores, all of which occurred within the paper-based iteration of our system (since the mandatory fields allowed us to prevent such missing data in the electronic version). Quality Assessment We used a novel evaluation tool [14] to determine the quality of the assessment comments. The derivation of that tool is described elsewhere [14]. Essentially, three authors (TC, SS, and SM) scored different subsets of the comments. In a previous study, this scoring process was calibrated using 10% of the dataset, achieving an intraclass correlation coefficient (ICC) of 0.95 [14]. Quantitative analysis All data were processed and coded using Microsoft Excel 2011 for Mac Descriptive Statistics; Pearson correlation, Chi-squared, and univariate analysis of variance (ANOVA) were calculated using IBM SPSS 26 (IBM Corp. Released 2019, IBM SPSS Statistics for Windows, Version 26.0. Armonk, NY: IBM Corp). To evaluate the impact of medium and years in practice on the three dependent variables, word count, QuAL score, and McMAP rating scores were analyzed separately. Pearson correlation analyses were used to describe the relationship between our dependent variables (word count, McMAP score, and QuAL score) and years in practice. Separate univariate ANOVAs were used to describe the impact of medium (paper or electronic) on our dependent variables (word count, McMAP score, and QuAL score). Chi-squared analysis was used to evaluate the independence of frequencies for records with three words or less between paper and electronic entries. Ethics Our project was reviewed by our local institutional review board’s chairperson (Hamilton Integrated Research Ethics Board) and granted an exemption.

Results

From October 2012 to July 2014, 2,556 assessments were generated using written comments, rating scale scores, or both. A total of 1,018 of these assessments were generated within paper workbooks and subsequently transcribed by an administrative staff member or junior faculty member. Of those 1,018 paper-based records, 33% (n=338) were excluded for having three words or less and 22% (n=225) records had no words. From July 2013 to July 2015, daily evaluations of both tasks and overall performance were collected via a novel electronic platform, yielding 1,538 assessments, which included both written comments and numerical ratings. Of those 1,538 records, 24% (n=373) were excluded for having three words or less and 16% (n=241) had zero words. Table 1 shows the distribution of records by year and medium. A Chi-squared analysis of the test of independence of the distribution of comments with less than three words was not significant (χ2=2.2, p=0.5). Assessments were from a total of 86 faculty members of whom 64% were male. Around 1% (20) of entries that met the inclusion criteria did not provide information about years in independent practice or rater identity; these records were excluded from the analysis. Table 2 shows the overall demographics of the assessors.
Table 1

Yearly distribution of comments with five words or greater

n=1,825.

YearElectronicPaper
20153020
20147220
2013136573
2012092
Table 2

Participant demographics and dataset details

 n%
Gender of trainees
Female1446.7
Male1653.3
Gender of raters
Female 3136.0
Male5564.0
Years in practice  
>201013.9
11 to 201820.9
≤10 yrs (including senior trainees)5867.4
Assessment type  
Paper66536.4
Electronic1,16063.6

Yearly distribution of comments with five words or greater

n=1,825. Word count After the exclusion of 731 records, the average word count was 15 (SD=14) across 1,825 comments. There was a significant main effect of medium on word count (F (1,1823)=52.87, p=0.01, partial η2=0.83). The electronic records had a higher word count (16) on average compared to paper-based records (12). Word count and years in practice Years in practice was negatively correlated to word count (r=-0.2; p<0.001). The correlation was the same when examining paper records and electronic records (for paper: r=-0.25, p<0.001; for electronic r=-0.19, p<0.001). Quality assessment of comments The average QuAL score, 0.47/5 (SD=0.86), was lower compared to the sample studied previously (mean=0.9/5, SD=0.9) [14]. The lower range of scores is consistent with prior work, indicating the continued need for faculty development when providing written feedback [14]. Evaluating the transition from paper to electronic data capture, the univariate ANOVA showed a main effect of medium (F (1, 1823)=6.7, p<0.01, partial η2=0.004). Critically, for our study goal, there was a small effect size, and the quality of comments was not reduced because of the transition, as the average QuAL score for electronically captured assessments was 0.51 compared to 0.41 for paper-based assessments. Unsurprisingly, longer comments (regardless of the medium) were positively correlated with scores on the first QuAL subscale (evidence of observed behavior; r=0.46, p<0.01), positively correlated with scores on the second subscale (suggestion for improvement; r=0.40, p<0.01), and positively correlated with scores on the third subscale (evidence linked to suggestion; r=0.41, p<0.001). Quality assessment of comments and years in practice Years in practice was negatively correlated with QuAL score (r=-0.08, p<0.001). McMAP global rating score The mean McMAP score was 5.9/7 (SD=0.82), with a median score of 6. There was no main effect of medium on McMAP scores. McMAP global rating score and years in practice There was no significant correlation between years in practice and McMAP global rating score. Although Figure 1 depicts some variability in scores across different years in practice, we did not detect a significant or meaningful relationship between years in practice and McMAP scores.
Figure 1

Mean McMAP scores of trainees generated in both contexts (paper vs. electronic)

Year 0 on the x-axis denotes the years since the end of their postgraduate training. Numbers less than zero (e.g., negative) on this axis denotes the number of years prelicensure for an individual since senior residents often acted as assessors for junior residents (e.g., senior emergency medicine resident would observe, provide feedback, and rate a first- or second-year trainee).

Mean McMAP scores of trainees generated in both contexts (paper vs. electronic)

Year 0 on the x-axis denotes the years since the end of their postgraduate training. Numbers less than zero (e.g., negative) on this axis denotes the number of years prelicensure for an individual since senior residents often acted as assessors for junior residents (e.g., senior emergency medicine resident would observe, provide feedback, and rate a first- or second-year trainee).

Discussion

Our study examined the relationship between the method of collecting assessment data and the quality and quantity of words within the comments for one WBA system. You will recall that our intention was to determine if the quality of comments generated in paper vs. electronic media was influenced by an assessor’s seniority. Contrary to the postulations of our residents in our previous qualitative program evaluation study [8], our faculty members were not deterred by the transition to electronic media and, on average, wrote more words for qualitative comments. Faculty members did skip comment boxes more often in the electronic version (23% vs. 9.2%), despite that there were mandatory fields within the digital version. Going beyond the initial mandate of our study, we also elucidated an interesting finding with regards to the volume of feedback generated by different cohorts of our attending physicians. Mid-career faculty members tended to write the least. In our locale, we hypothesize that the phenomenon we observed may be due to the effects described by Govaerts and colleagues [13], but the phenomenon may also intersect with faculty engagement. In our local quality assurance focus groups, residents revealed that many faculty members were rather disengaged with the new WBA system (McMAP) [8]. As such, we postulate that a number of different forces may be at play. With the advent of competency-based medical education (CBME), there has been a marked use of digital systems to capture WBA [19-22]. With the increasing use of these databases, many groups have resorted to trainee behaviors around data capture [23,24]. However, more attention must be paid to how faculty respond and engage with these systems and then how faculty respond to their needs via faculty development [25]. Whereas in traditional testing, validity lies in the hands of the students and their engagement in the response process, in the age of CBME and WBAs, the response process of faculty members who enter the data is of paramount importance. Our present study sheds light on an important aspect of the response process for generating high-quality data about trainee performance. Others have examined time burden on faculty [26] with WBA, engagement of faculty [27] in the assessment process, biases they exhibit [28], and even perceptions of their role within these systems [16,17]. In our study, by examining the WBA participation of various faculty cohorts, we show how we might bring more nuanced analyses around different needs of various subgroups of faculty. By doing this type of analysis, we feel that we could begin to refine approaches for faculty development. Rather than seeing faculty as one singular group, more nuanced and targeted approaches to faculty development can be generated by transforming trainee databases to reveal new insights about faculty performance [29]. Next Steps Trainee databases and repositories may represent a wealth of untapped data that can provide faculty with tangible, actionable insights about their own performance as faculty raters within a system of assessment. A true digital transformation of faculty development may be possible if we harness the newly developed trainee assessment databases to generate useful metrics on faculty performance in terms of their contributions to assessment, feedback, and rating of trainees in the age of CBME [30]. Repurposing trainee data for faculty development insights holds great potential for providing true insights into actual faculty performance related to assessment and their tangible contributions to academic medicine. Future studies in this area may include studies that examine sentiment analysis or applications of natural language processing (such as sequencing of feedback statements, syntactic complexity, local or text coherence, lexical sophistication) to the real-time data capture of trainee feedback comments. Limitations This study has a number of limitations. This is a retrospective program evaluation study. Novelty effects of technology may have distorted the use of electronic vs. paper. However, the increased use of the electronic medium can reduce the technological barrier for the raters and create better buy-in to generate feedback to residents. The paper version of data was only from year one, so the score changes from year to year may reflect increasing score drift due to increased usage by faculty members. Therefore, we cannot tease apart the initial pilot year’s novelty effect on our present study. Finally, the data are coming from a single institution with a focus on emergency medicine. The generalizability of our results is limited to our research population.

Conclusions

We detail our journey through the effective digitalization of a WBA system, which resulted in more words written per comment about trainee performance, and the presence of higher-quality comments. True digital transformation may be possible by harnessing trainee data repositories and repurposing them to analyze for faculty-relevant metrics. Digitalization of workplace-based assessments resulted in an increase in the length of comments available to educators. Longer comments achieved on digital systems did not appear to negatively impact the quality of the assessments. The medium, electronic vs. paper, has a high influence on the evaluative message being sent or received.
  25 in total

1.  Using natural language processing to provide personalized learning opportunities from trainee clinical notes.

Authors:  Joshua C Denny; Anderson Spickard; Peter J Speltz; Renee Porier; Donna E Rosenstiel; James S Powers
Journal:  J Biomed Inform       Date:  2015-06-10       Impact factor: 6.317

2.  The McMaster Modular Assessment Program (McMAP): A Theoretically Grounded Work-Based Assessment System for an Emergency Medicine Residency Program.

Authors:  Teresa Chan; Jonathan Sherbino
Journal:  Acad Med       Date:  2015-07       Impact factor: 6.893

Review 3.  Advancing Workplace-Based Assessment in Psychiatric Education: Key Design and Implementation Issues.

Authors:  John Q Young; Jason R Frank; Eric S Holmboe
Journal:  Psychiatr Clin North Am       Date:  2021-06

4.  From Utopia Through Dystopia: Charting a Course for Learning Analytics in Competency-Based Medical Education.

Authors:  Brent Thoma; Rachel H Ellaway; Teresa M Chan
Journal:  Acad Med       Date:  2021-07-01       Impact factor: 6.893

5.  Wresting with Implementation: a Step-By-Step Guide to Implementing Entrustable Professional Activities (EPAs) in Psychiatry Residency Programs.

Authors:  Erick K Hung; Michael Jibson; Julie Sadhu; Colin Stewart; Ashley Walker; Lora Wichser; John Q Young
Journal:  Acad Psychiatry       Date:  2020-10-20

6.  Beyond summative decision making: Illuminating the broader roles of competence committees.

Authors:  Rachael Pack; Lorelei Lingard; Christopher Watling; Sayra Cristancho
Journal:  Med Educ       Date:  2020-04-01       Impact factor: 6.251

7.  Evaluation of a National Competency-Based Assessment System in Emergency Medicine: A CanDREAM Study.

Authors:  Brent Thoma; Andrew K Hall; Kevin Clark; Nazanin Meshkat; Warren J Cheung; Pierre Desaulniers; Cheryl Ffrench; Allison Meiwald; Christine Meyers; Catherine Patocka; Lorri Beatty; Teresa M Chan
Journal:  J Grad Med Educ       Date:  2020-08

8.  Workplace-based assessment: effects of rater expertise.

Authors:  M J B Govaerts; L W T Schuwirth; C P M Van der Vleuten; A M M Muijtjens
Journal:  Adv Health Sci Educ Theory Pract       Date:  2010-09-30       Impact factor: 3.853

9.  A mobile app to capture EPA assessment data: Utilizing the consolidated framework for implementation research to identify enablers and barriers to engagement.

Authors:  John Q Young; Rebekah Sugarman; Jessica Schwartz; Matthew McClure; Patricia S O'Sullivan
Journal:  Perspect Med Educ       Date:  2020-08

10.  Developing a dashboard for faculty development in competency-based training programs: a design-based research project.

Authors:  Yusuf Yilmaz; Robert Carey; Teresa M Chan; Venkat Bandi; Shisong Wang; Robert A Woods; Debajyoti Mondal; Brent Thoma
Journal:  Can Med Educ J       Date:  2021-09-14
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.