Literature DB >> 26316836

Cross-cultural challenges for assessing medical professionalism among clerkship physicians in a Middle Eastern country (Bahrain): feasibility and psychometric properties of multisource feedback.

Ahmed Al Ansari1, Khalid Al Khalifa2, Mohamed Al Azzawi2, Rashed Al Amer2, Dana Al Sharqi3, Anwar Al-Mansoor4, Fadi M Munshi5.   

Abstract

BACKGROUND: We aimed to design, implement, and evaluate the feasibility and reliability of a multisource feedback (MSF) system to assess interns in their clerkship year in the Middle Eastern culture, the Kingdom of Bahrain.
METHOD: The study was undertaken in the Bahrain Defense Force Hospital, a military teaching hospital in the Kingdom of Bahrain. A total of 21 interns (who represent the total population of the interns for the given year) were assessed in this study. All of the interns were rotating through our hospital during their year-long clerkship rotation. The study sample consisted of nine males and 12 females. Each participating intern was evaluated by three groups of raters, eight medical intern colleagues, eight senior medical colleagues, and eight coworkers from different departments.
RESULTS: A total of 21 interns (nine males and 12 females) were assessed in this study. The total mean response rates were 62.3%. A factor analysis was conducted that found that the data on the questionnaire grouped into three factors that counted for 76.4% of the total variance. These three factors were labeled as professionalism, collaboration, and communication. Reliability analysis indicated that the full instrument scale had high internal consistency (Cronbach's α 0.98). The generalizability coefficients for the surveys were estimated to be 0.78.
CONCLUSION: Based on our results and analysis, we conclude that the MSF tool we used on the interns rotating in their clerkship year within our Middle Eastern culture provides an effective method of evaluation because it offers a reliable, valid, and feasible process.

Entities:  

Keywords:  MSF system; generalizability; interns; validity

Year:  2015        PMID: 26316836      PMCID: PMC4540155          DOI: 10.2147/AMEP.S86068

Source DB:  PubMed          Journal:  Adv Med Educ Pract        ISSN: 1179-7258


Introduction

The regular evaluation of medical professionals is an essential step in the process toward improving the quality of medical care provided by any health care institution.1 Such evaluations point to both the strengths and weaknesses found amongst the health care staff.2 Thorough evaluation of medical interns is critical as this is the starting point of their professional careers. The feedback received from these evaluations will address the areas of improvement they might need to address with more attention.3 Multisource feedback (MSF), also called the 360-degree evaluation, is an evaluation process by which various clinicians fill out surveys to evaluate their medical peers and colleagues. This provides feedback from individuals other than the attending and supervising physicians.4 This type of assessment uses raters from a variety of groups who interact with trainees. Though this type of evaluation process may seem subjective,5 several studies have showed that this type of assessment is reliable and valid provided that enough assessors are used to eliminate the bias factor.6 MSF is advocated as a feasible approach to evaluating physicians on their interpersonal relationships, and interactions in particular.7 MSF is gaining acceptance and credibility as a means of providing doctors with relevant information about their practice, to help them monitor, develop, maintain, and improve their competence. MSF is focused on clinical skills, communication, and collaboration with other health care professionals, professionalism, and patient management.8 As such, it has been argued that all physician evaluation processes include MSF as a key component.9 Although MSF has been applied across many clinical settings using a variety of different medical specialties, this type of assessment has not been used in graduate medical education, for those rotating in the clerkship year, nor has it been used in the Middle Eastern culture. We used MSF to evaluate the medical interns at Bahrain Defense Force (BDF) Hospital. The goal of this study was to assess the feasibility, reliability, and validity of the MSF process for the clerkship year specifically and in the Middle East generally.

Methods

Study settings and participants

This study was conducted in the BDF hospital, a military teaching hospital in the Kingdom of Bahrain. The BDF hospital is a military hospital with 450 beds, 322 physicians and dentists, and 1,072 nurses and practical nurses. Annually, 32,462 patients are seen as inpatients, along with over 347,025 outpatients seen in the clinics. The internship year is a full academic year that undergraduate medical school students must complete after finishing medical school. The internship year program for the study year class consisted of the following rotations: 3 months in Surgery, 3 months in Internal Medicine, 2 months in Pediatrics, 2 months in Obstetrics and Gynecology, 1 month in Emergency Medicine, and 1 month an elective of the student’s choice. A total 21 interns (who represented the total population of the interns for the year), who had finished their medical school programs and started their 1-year clerkship rotation in our military teaching hospital were assessed in this study. The study included nine males and 12 females. Sixteen interns had graduated from the Royal College of Surgeons of Ireland (RCSI) in Bahrain, two had graduated from medical schools in Egypt, one from Sudan, one from Saudi Arabia, and one had graduated from Yemen.

Raters

Research has shown that when physicians choose their own raters, the resulting evaluations are not significantly different from evaluations by raters selected by a third party.10 Furthermore, studies evaluating how familiar the physician was with his/her rater found that familiarity did not alter the ratings significantly.11 Despite these findings and due to the cultural differences in the Middle East, we hypothesized that the self-selection of raters may influence the results of the evaluations; therefore, we decided in this study to select raters randomly for each intern. The basic criterion was that the intern must know and have worked with his/her individual raters for a minimum of 2 months. Different groups of raters were selected for this study. Each intern was evaluated by three groups of raters: eight medical intern colleagues, eight senior medical colleagues (such as chief residents and consultants from different departments), and eight coworkers from different departments. An independent administrative team was formed to carry out the responsibility of distributing the evaluation materials and collecting them in sealed envelopes. Each envelope consisted of the evaluation instrument and a guide for the implementation of the MSF and the purpose of the study. As this was the first time that MSF has been implemented in the hospital, it was necessary to clarify several points for the raters, as some of these points may be essential for their accurate rating of interns. We clarified in the guideline form that the results would be used to improve the interns’ performance and not for their promotion or selection into the residency training programs. Also, the main purpose of this research was stated as: to assess the feasibility, reliability, and validity of implementing the MSF system in the organization. It was also described that each participating intern would receive feedback from his/her evaluations to point to individual areas that needed improvement. Each evaluator was sent the forms that he/she was required to complete. Two weeks later, an e-mail was sent as a reminder from the administrative team. After another 2 weeks, raters who did not submit their envelopes received a phone call from the administrative team as a second reminder.

Statistical analysis

A number of statistical analyses were undertaken to address the research questions posed. Response rates and the time required to fill out the questionnaire were used to determine feasibility for each of the respondent groups.7,12 For each item on the survey, the percentage of the “unable-to-assess” answers, along with the mean and standard deviation, was computed to determine the viability of the items and the score profiles. Items for which the “unable-to-assess” answer exceeded 20% on a survey might be in need of revision or deletion according to previously conducted research.12 We used exploratory factor analysis to determine which items on each survey were suitable to group together (ie, become a factor or scale). In this study, using individual-physician data as the unit of analysis for the survey, the items were intercorrelated using Pearson product moment correlations. The correlation matrix was then broken down into its main components, and these components were then rotated to the normalized varimax criterion. Items were considered to be part of a factor if their primary loading was on that factor. The number of factors to be extracted was based on the Kaiser rule (ie, eigenvalues >1.0).13 The factors or scales established through exploratory factor analysis were used to establish the key domains for improvement (eg, professionalism), whereas the items within each factor provided more precise information about specific behaviors (eg, maintains confidentiality of patients, recognizes boundaries when dealing with other physicians, and shows professional and ethical behavior). Physician improvement could be guided by the scores on factors or items. This analysis made it possible to determine whether the instrument items were aligned into the appropriate constructs (factors) as intended. Instrument reliability (stability) was assessed. Internal consistency/reliability was examined by calculating the Cronbach’s coefficient for each of the scales and for each factor separately. Cronbach’s coefficient is widely used to evaluate the overall internal consistency for each instrument as well as for the individual factors within the instruments.14 This analysis was followed by a generalizability analysis to determine the generalizability coefficient (Ep2) and to ensure there were ample numbers of questions and evaluators to provide accurate and stable results for each intern on each instrument. Normally, an Ep2 =0.70 suggests data are stable. If the Ep2 is below 0.70, it suggests that more raters or more items are required to enhance stability. The G analyses were based on a single-facet, nested design with raters nested within the doctors who were being assessed, using the formula:15 where “var comp” is the component variability. Although this type of design does not allow for estimation of the interaction effect of raters with the doctors they are evaluating, it does allow for determination of the Ep2 of raters. We further conducted a D study where we estimated the Ep2 for one to ten raters.

Ethical approval

The research was approved by the research ethics committee of the BDF hospital. Written consent was obtained from the interns and verbal consent was obtained from raters. The study was conducted from January 2014 to May 2014.

Instrument

A modified instrument was developed based on the Physician Achievement Review (PAR) instrument, which was used to assess physicians in Alberta.16 The focus of the instrument is the assessment of professionalism, communication skills, and collaboration. To establish face and content validity, a table of specifications was constructed, and a working group was involved in developing the instrument. Expert opinion was taken into consideration as well. The instrument consisted of 39 items: 15 items to assess professionalism, 13 items to assess communication skills, and eleven items to assess collaboration. The instrument was constructed in a way that can be applied by different groups of people, including interns, senior medical colleagues, consultants, and coworkers. The items on the instrument had a five-point Likert response scale, where 1= among the worst, 2= bottom half, 3= average, 4= top half, and 5= among the best, with the additional option of “unable-to-assess”. After the committee had developed the questionnaires, they were sent to every physician whose work fit the profile for episodic care, for feedback. The questionnaires were modified upon receipt of that feedback.

Results

A total of 21 interns, nine males and 12 females, who represented the total number of interns rotating in the BDF hospital for the year 2014, were assessed. A total of 16 interns had graduated from the RCSI in Bahrain, two from a medical school in Egypt, one from Sudan, one from Saudi Arabia, and one from Yemen. Each intern was evaluated by three different groups. Group 1 consisted of the medical intern colleagues, group 2 included the senior medical colleagues (chief residents/consultants), and group 3 comprised coworkers. The total number of collected forms was 314, including 105 surveys from coworkers, 93 surveys from medical intern colleagues, and 116 surveys from senior medical colleagues (chief residents/consultants). Each intern was evaluated by different number of forms, ranging between 13 and 19 forms for each intern. The total mean response rates were 62.3%, and the time needed to complete each questionnaire was 7 minutes, which shows the feasibility of the survey.7,12 Most of the questions were answered by the respondents. There were only four questions (27, 28, 36, and 38) that exceeded 20% for the response “unable-to-assess” by the raters. Those questions may need to be reviewed, revised, or deleted in future implementation. However, even with the elimination of those four questions, the reliability of our survey remains high (Cronbach’s α =0.98). The whole instrument was found to be suitable for the factor analysis (Kaiser–Myer–Olkin [KMO] =0.953; Bartlett test significant, P<0.00). The factor analysis showed that the data on the questionnaire could be grouped into three factors that represented 76.4% of the total variance: professionalism, collaboration, and communication (Table 1).
Table 1

Descriptive statistics, item analysis, and factor analysis

QNMSD% UASelf (mean)SDFactors identified by factor analysis
CommCollaProfe
Q1Maintains confidentiality of patients2674.400.7312.14.550.660.78
Q2Recognizes boundaries when dealing with other physicians2854.380.769.204.500.790.82
Q3Recognizes boundaries when dealing with other health care professionals2854.400.759.204.750.620.79
Q4Shows professional and ethical behavior2914.460.777.304.770.440.68
Q5Is punctual, and performs tasks in a time-appropriate manner2834.270.889.804.580.670.66
Q6Is able to handle situations in a professional manner and exhibits self-control, avoiding emotional outbursts in stressful situations2734.160.8413.04.000.710.53
Q7Respects patient’s autonomy and right to be involved in his/her own management2764.390.7112.14.770.440.57
Q8Is reliable and responsible when performing his duties2884.300.868.284.770.560.72
Q9Is honest and handles his/her duties in a dignified manner2844.400.749.554.770.440.68
Q10Accepts constructive criticism and develops goals for improvement2654.130.8515.64.460.660.65
Q11Respects cultural, individual, and role differences including age, sex, race, religion, disability, language, sexual orientation, and socioeconomic status2814.510.6610.54.840.370.63
Q12Follows institutional policies and procedures2854.320.799.234.610.650.73
Q13Arrives on time to scheduled appointments and hospital activities2794.230.8814.34.530.660.77
Q14Manages health care resources efficiently2634.270.7315.94.330.650.49
Q15Leads with respect and fair treatment of colleagues2794.370.7211.14.610.510.56
Q16Communicates efficiently and in a clear, understandable fashion with colleagues within his/her team2854.400.759.234.600.760.64
Q17Communicates efficiently and in a clear, understandable, and compassionate way with patients2754.390.7612.44.530.770.69
Q18Allows the patient to elaborate about his condition2734.340.7616.24.380.760.72
Q19Communicates efficiently and in a clear, understandable, and compassionate way with patients’ families2614.280.8016.84.330.770.69
Q20Communicates clearly and effectively with other health care workers eg, nurses2814.330.7710.54.460.760.59
Q21Explains what is being done for the patient during examination or procedures2554.230.7718.74.380.870.68
Q22Communicates purpose and results of investigations to patients well2514.210.7812.04.100.830.73
Q23Follows up appropriately and in a timely manner on patients’ hospital course2594.220.8317.54.300.750.56
Q24Communicates management options to patients in a clear, understandable way, taking into account the patients’ opinions2644.240.7815.94.230.920.70
Q25Displays empathy in dealing with patients by eye contact and verbal responses2744.300.7812.74.460.760.57
Q26Summarizes the information given for the patient in small quantities, with concrete explanations and understandable language2524.230.7919.74.230.720.69
Q27Maintains calm in emergency situations, in order to communicate information clearly to his/her seniors2354.190.7825.14.100.790.60
Q28Communicates accurate patient information to physicians from other departments when required to do so2473.270.8021.34.230.830.69
Q29Manages to work well as part of a health care team2864.320.818.904.610.650.61
Q30Facilitates the learning of medical colleagues and coworkers2544.290.7919.14.390.650.68
Q31Collaborates well with nurses and other health care workers2804.320.818.904.530.770.65
Q32Concerned about the safety of patients and coworkers2694.510.8214.34.690.630.60
Q33Coordinates patient care efficiently2694.300.7714.24.610.660.61
Q34Collaborates with other health care workers in order to achieve optimal patient care2764.230.8412.14.690.630.65
Q35Participates in a system of call in order to provide care for patients2734.320.8313.04.750.620.65
Q36Provides appropriate guidance and help to team members on regular basis2474.240.8321.34.500.670.61
Q37Takes on extra work, when appropriate, to help the team2694.130.9314.34.460.660.69
Q38Enables the team to achieve agreements for team process and collaborative completion of assignment2474.220.8121.34.500.670.67
Q39Participates fully in collaborative process and fulfilled team agreements2564.220.8018.44.580.670.68

Abbreviations: Colla, collaboration; Comm, communication skills; M, mean; N, number of respondents; Profe, professionalism; Q, question; SD, standard deviation; Self, self-assessment; UA, unable to assess.

The reliability analysis (Cronbach’s α reliability of internal consistency) indicated that the instrument’s full scale had high internal consistency (Cronbach’s α =0.98). The reliability measure of the factors (subscales) within the questionnaire also had high internal consistency reliability (Cronbach’s α ≥0.91). G study analysis was conducted employing a single-facet, nested design. The Ep2 was 0.78 for the surveys. Also shown in Table 2 is a D study, where we estimated the Ep2 for one to ten raters: for one rater, Ep2 =0.30, for eight raters, Ep2 =0.78; and for ten raters, Ep2 =0.81.
Table 2

G study. The number of raters required and the generalizability coefficient

Number of ratersEp2
10.302
20.464
30.565
40.634
50.684
60.722
70.752
80.776
90.796
100.812

Note: Coefficients were based on a D study.

Abbreviation: Ep2, generalizability coefficient.

Discussion

In this study, we evaluated the applicability of questionnaire-based assessments, such as the MSF, of the interns who rotated in our military teaching hospital during their clerkship year. To our knowledge, this was the first study to combine feedback from medical intern colleagues, senior medical colleagues, such as chief residents, consultants, and coworkers, and from medical support staff for the assessment of interns in the clerkship year. In addition, we believe that this was the first study investigating the feasibility, reliability, and validity of the MSF system in the Middle East. In this study, a set of MSF questionnaires to assess interns (by fellow interns, senior medical colleagues, and coworkers) was developed and evaluated. Also, the aim was to assess the feasibility and reliability of the instruments and to explore the evidence of validity. Interns were assessed on a number of aspects of practice that the regulatory authority and the physicians themselves believed to be important. Although not designed to specifically assess Accreditation Council for Graduate Medical Education or CanMEDS competencies, the items and factors used allowed us to evaluate some aspects of both sets of competencies, respectively.17 However, to develop a tool that fully assessed either set of competencies would have required the addition of new items and the retesting of the instrument, as well as its factors. This type of assessment is feasible in the contextual setting described because this was demonstrated by the adequate response rates. Although this may be partly explained by the fact that this type of assessment is a new form of evaluation for both the BDF hospital and Middle Eastern culture, the received response rates were consistent with the response rates that were found in other studies.18,19 Initial evidence for the validity of the instruments has been found, recognizing that establishing validity is a process that cannot be proven from a one-time study. Almost all of the questions were answered by the responding physicians and coworkers. However, there were items on all questionnaires that many of the respondents were unable to assess; these items must be reviewed. Some may be amenable to modification, while others may need to be deleted. Our exploratory factor analyses showed that items did, in fact, group together into factors in the same way as predicted by the table of specifications. Regulatory authorities tend to focus on improving both professionalism and collaboration, especially amongst young physicians.20 As such, the factors identified in our results provided a larger picture of the areas of improvement for our physician population as a whole, whereas the individual items used pointed to more specific feedback. Each physician received descriptive data (both means and standard deviations) on the scales and individual items, for him/herself as well as for the group as a whole. Additional work will be required to examine the validity of the instruments. For example, it would be useful to determine whether physicians who received high scores on the MSF evaluation are also high performers on other assessments that examine performance more objectively. Similar work to identify the criterion validity was examined by other researchers in different settings. Yang et al found a medium correlation between the 360 degree evaluation and the small scale Objective Structured Clinical Examination (OSCE) (r=0.37, P<0.05). This finding supports the criterion validity of the MSF.21 There are limitations to this study. This study focused on interns in one hospital in the Kingdom of Bahrain. We cannot currently conclude whether our results would be applicable to different interns in other countries in the Middle East. Another limitation of the current study was the limited number of participants, using 21 interns in total.

Conclusion and future research

Determining the validity and the reliability of the MSF system in assessing interns was a challenge in our institution. The MSF instruments for interns provide a feasible way of assessing interns in their clerkship year and of providing guided feedback on numerous competencies and behaviors. Based on the present results, the currently used instruments and procedures have high reliability, validity, and feasibility. The item analyses, reliability, and factor analyses all indicate that the present instruments are working generally very well in assessing interns rotating through the hospital in their clerkship year. Replicating this study in other Middle Eastern settings may provide evidence to support the validity of the MSF process.
  19 in total

1.  Changing physicians' practices: the effect of individual feedback.

Authors:  H Fidler; J M Lockyer; J Toews; C Violato
Journal:  Acad Med       Date:  1999-06       Impact factor: 6.893

2.  Setting performance standards for medical practice: a theoretical framework.

Authors:  L Southgate; R B Hays; J Norcini; H Mulholland; B Ayers; J Woolliscroft; M Cusimano; P McAvoy; M Ainsworth; S Haist; M Campbell
Journal:  Med Educ       Date:  2001-05       Impact factor: 6.251

Review 3.  Multisource feedback in the assessment of physician competencies.

Authors:  Jocelyn Lockyer
Journal:  J Contin Educ Health Prof       Date:  2003       Impact factor: 1.355

4.  Use of SPRAT for peer review of paediatricians in training.

Authors:  Julian C Archer; John Norcini; Helena A Davies
Journal:  BMJ       Date:  2005-05-09

5.  Self and peer assessment of pediatricians, psychiatrists and medicine specialists: implications for self-directed learning.

Authors:  Claudio Violato; Jocelyn Lockyer
Journal:  Adv Health Sci Educ Theory Pract       Date:  2006-08       Impact factor: 3.853

6.  The CanMEDS initiative: implementing an outcomes-based framework of physician competencies.

Authors:  Jason R Frank; Deborah Danoff
Journal:  Med Teach       Date:  2007-09       Impact factor: 3.650

7.  Use of 360-degree assessment of residents in internal medicine in a Danish setting: a feasibility study.

Authors:  P Allerup; K Aspegren; E Ejlersen; G Jørgensen; A Malchow-Møller; M K Møller; K K Pedersen; O B Rasmussen; A Rohold; B Sørensen
Journal:  Med Teach       Date:  2007-03       Impact factor: 3.650

8.  Assessment of psychiatrists in practice through multisource feedback.

Authors:  Claudio Violato; Jocelyn M Lockyer; Herta Fidler
Journal:  Can J Psychiatry       Date:  2008-08       Impact factor: 4.356

9.  Use of peer ratings to evaluate physician performance.

Authors:  P G Ramsey; M D Wenrich; J D Carline; T S Inui; E B Larson; J P LoGerfo
Journal:  JAMA       Date:  1993-04-07       Impact factor: 56.272

10.  The assessment of emergency physicians by a regulatory authority.

Authors:  Jocelyn M Lockyer; Claudio Violato; Herta Fidler
Journal:  Acad Emerg Med       Date:  2006-11-10       Impact factor: 3.451

View more
  4 in total

1.  A Systematic Review of the Quality and Utility of Observer-Based Instruments for Assessing Medical Professionalism.

Authors:  Yu Heng Kwan; Kelly Png; Jie Kie Phang; Ying Ying Leung; Hendra Goh; Yi Seah; Julian Thumboo; A/P Swee Cheng Ng; Warren Fong; Desiree Lie
Journal:  J Grad Med Educ       Date:  2018-12

Review 2.  Assessing medical professionalism: A systematic review of instruments and their measurement properties.

Authors:  Honghe Li; Ning Ding; Yuanyuan Zhang; Yang Liu; Deliang Wen
Journal:  PLoS One       Date:  2017-05-12       Impact factor: 3.240

3.  Selecting graduates for the interns' award by using multisource feedback process: does it work?

Authors:  Kathryn Strachan; Sameer Otoom; Amal Al-Gallaf; Ahmed Al Ansari
Journal:  BMC Res Notes       Date:  2017-10-30

4.  Intercultural Competence of Western Teachers for Nepalese Rescuers.

Authors:  Monika Brodmann Maeder; Raphael Saghir; Matiram Pun; Agnieszka Elzbieta Stawinoga; Rachel Turner; Giacomo Strapazzon; Aristomenis K Exadaktylos; Hermann Brugger
Journal:  High Alt Med Biol       Date:  2018-12-08       Impact factor: 1.981

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.