Literature DB >> 19188458

Application of quality improvement strategies in 389 European hospitals: results of the MARQuIS project.

M J M H Lombarts1, I Rupp, P Vallejo, R Suñol, N S Klazinga.   

Abstract

CONTEXT: This study was part of the Methods of Assessing Response to Quality Improvement Strategies (MARQuIS) research project investigating the impact of quality improvement strategies on hospital care in various countries of the European Union (EU), in relation to specific needs of cross-border patients. AIM: This paper describes how EU hospitals have applied seven quality improvement strategies previously defined by the MARQuIS study: organisational quality management programmes; systems for obtaining patients' views; patient safety systems; audit and internal assessment of clinical standards; clinical and practice guidelines; performance indicators; and external assessment.
METHODS: A web-based questionnaire was used to survey acute care hospitals in eight EU countries. The reported findings were later validated via on-site survey and site visits in a sample of the participating hospitals. Data collection took place from April to August 2006.
RESULTS: 389 hospitals participated in the survey; response rates varied per country. All seven quality improvement strategies were widely used in European countries. Activities related to external assessment were the most broadly applied across Europe, and activities related to patient involvement were the least widely implemented. No one country implemented all quality strategies at all hospitals. There were no differences between participating hospitals in western and eastern European countries regarding the application of quality improvement strategies.
CONCLUSIONS: Implementation varied per country and per quality improvement strategy, leaving considerable scope for progress in quality improvements. The results may contribute to benchmarking activities in European countries, and point to further areas of research to explore the relationship between the application of quality improvement strategies and actual hospital performance.

Entities:  

Mesh:

Year:  2009        PMID: 19188458      PMCID: PMC3269892          DOI: 10.1136/qshc.2008.029363

Source DB:  PubMed          Journal:  Qual Saf Health Care        ISSN: 1475-3898


Quality and safety of patient care are high on the European policy agenda, as evidenced by various commitments by European health ministries. Patient mobility has clearly been a triggering factor. Governments may fear that differences in the perceived quality or costs of health services may encourage patients to cross borders to obtain healthcare.1–3 The fact that member states are now talking about what is still their responsibility has increased the need for information about the cross-border movement. The Methods of Assessing Response to Quality Improvement Strategies (MARQuIS) research project aims to be instrumental in providing a better understanding of this movement, by investigating and comparing different quality improvement (QI) strategies in healthcare systems across the European Union. In this article we focus on the degree to which QI strategies are applied at European hospitals, by their own report. As presented in the Health Care Quality Strategies in Europe study, we identified seven national QI strategies.4 5 Our primary focus is on implementation of the strategies at the EU level; data at the country level are reported for reference purposes. The QI strategies we investigated were:The countries participating in this study were Spain, France, Poland, Czech Republic, the UK, Ireland, Belgium and the Netherlands. organisational quality management programmes; systems for obtaining patients’ views; patient safety systems; audit and internal assessment of clinical standards; clinical and practice guidelines; performance indicators and measurements; external assessment.

MATERIAL AND METHODS

Questionnaire design

We conducted a web-based questionnaire survey. The questionnaire was developed to measure QI, defined as the application of quality policies and procedures, quality governance structures, and quality activities used to close the gap between current and expected levels of quality.4 To determine the distinctive aspects of QI we used several sources, such as existing QI questionnaires,6–9 a review of the quality literature,10–12 an analysis of accreditation manuals,11 13 14 and the results of previous MARQuIS studies including a literature review covering QI strategies in member states of the EU, and an analytical framework defining areas of QI policies and strategies. A glossary of quality concepts and tools was made available to participants. The questionnaire consisted of four sections: one section focused on QI at the hospital level, the other three on quality management for specific medical conditions. The three medical conditions for focused data collection were selected based on two criteria: the condition had to represent a significant volume of cross-border patient care,15 and the combination of conditions was intended to cover the most relevant services offered by a hospital—that is, emergency surgical and medical services, and maternal and neonatal services. The three conditions selected were acute myocardial infarctions (AMI), acute appendicitis and deliveries. For each condition selected, the literature was searched for specific QI strategies. Search terms used included “quality assurance”, “quality improvement”, “quality assessment”, and “performance measurement”.16–31 We stopped searching when additional publications no longer resulted in new relevant QI strategies, activities or measures. Practising medical specialists were consulted for their comments and suggestions on the specific QI strategies, activities, and measures (see Acknowledgements). Both members of the MARQuIS team and the nine country coordinators reviewed the draft questionnaire. (For Belgium two country coordinators were appointed, one for Flemish-speaking and one for French-speaking hospitals.) The questionnaire was then pilot tested in two hospitals in Ireland and the UK (these countries were chosen for language-related reasons), and a few amendments were made as a result. The questionnaire was translated into five languages (Spanish, French, Polish, Czech, and Dutch); the country coordinators were responsible for translation. A forward–backward translation protocol was used. Figure 1 shows the final structure of the questionnaire, which totalled 199 questions. For each of the four sections a preferred respondent (at the senior level) was suggested.
Figure 1

Structure of the MARQuIS questionnaire. AMI, acute myocardial infarction; QI, quality improvement.

Response categories

Various scoring scales were used depending on the type of question. Items were scored on a two-point scale (yes/no), a four-point scale (see tables 2 to 8 in Results), or a five-point Likert scale (1 =  strongly agree, 5 =  strongly disagree).
Table 2

Quality improvement (QI) strategies as applied in European hospitals: organisational quality management programmes/activities; total and per country*, numbers are positive responses in valid percentages (total item response in absolute numbers)

Specification of QI strategyTotalIrelandBelgiumFranceSpainPolandCzech Republic
The use of ISO in implementing a quality system51.2 (336)31.6 (19)21.7 (23)28.6 (63)70.3 (101)50.0 (76)59.5 (37)
The use of EFQM in implementing a quality system29.5 (319)20.0 (20)60.9 (23)3.4 (59)57.4 (101)4.3 (69)20.0 (30)
Active quality improvement team(s)/circles†3492324651067637
    127.547.84.27.719.842.148.6
    217.517.441.720.016.014.513.5
    338.130.429.247.753.823.713.5
    413.24.325.023.18.59.221.6
Committee or person responsible for:
    Hospital infections control98.9 (353)100 (25)100 (24)100 (78)100 (107)100 (80)89.5 (38)
    Blood transfusion92.2 (348)100 (25)100 (24)100 (78)87.5 (104)92.1 (76)78.9 (38)
    Prevention of bed sores85.3 (346)66.7 (21)95.8 (24)79.7 (64)94.2 (104)78.9 (76)84.6 (39)
    Policy on use of antibiotics92.6 (352)87.0 (23)95.8 (24)96.9 (65)98.1 (106)96.1 (76)66.7 (39)

*The results for the UK and the Netherlands are included in the total scores but not listed separately, due to the very low response rates.

†The sum of percentages may not always equal 100%; the percentage answers “don’t know/no answer” are not listed here. 1 =  yes, in most departments (>50%); 2 =  yes, in most departments (>50%), but not systematically; 3 =  yes, in some departments (<50%); 4 =  no.

EFQM, European Foundation for Quality Management.

*The results for the UK and the Netherlands are included in the total scores but not listed separately, due to the very low response rates. †The sum of percentages may not always equal 100%; the percentage answers “don’t know/no answer” are not listed here. 1 =  yes, in most departments (>50%); 2 =  yes, in most departments (>50%), but not systematically; 3 =  yes, in some departments (<50%); 4 =  no. EFQM, European Foundation for Quality Management.

Sampling and recruitment

Our survey focused on European hospitals with a minimum of 100 acute care beds, and offering care for at least two out of the three conditions selected for study (AMI, appendicitis and deliveries). Two additional criteria defined the hospital sample: ownership status (public, private not-for-profit and private for-profit), and actual or potential cross-border care delivery.31 The target was to include a total of 600 hospitals covering eight countries. For smaller countries (Belgium, the Netherlands, Ireland, Czech Republic) this meant that all hospitals meeting the inclusion criteria were invited to participate. In the remaining countries, hospitals were randomly sampled from a list of hospitals that met the sampling criteria. Hospital recruitment was done by the country coordinators, who used different strategies. To make participation more appealing to hospitals, a package of advantages was offered to the hospitals, including membership in the MARQuIS network, a certificate of participation, and free subscription to the project’s newsletter. Hospitals that agreed to participate in the survey received an e-mail inviting them to enter the MARQuIS website (http://www.marquis.be) and fill out the web-based questionnaire. Upon request, a paper version of the questionnaire could be used and sent to the researchers (MJMHL or IR). Data were collected from the beginning of April to the end of August 2006. Hospitals received up to three reminders. Again, country coordinators used different approaches to raise the response rates.

Validation of the data

To validate the questionnaire data, two analyses were performed by using data collected during on-site hospital visits in a selected sample of 89 hospitals that had previously completed the questionnaire. Visits were performed by independent external auditors. All aspects of on-site visits are described in detail elsewhere.32 During the visits, the hospitals’ key informants were asked to answer 25 questions that had been previously asked in the questionnaire. The reliability of the questionnaire was assessed by the level of agreement between the responses to these 25 questions as given in the questionnaire and during the on-site visits. In addition, for 14 of these 25 questions the external auditors requested evidence to check the answers given during the audit. Criterion validity was then assessed as the degree of agreement between the information provided by key informants and the evidence found to underpin this information. Reliability and validity were assessed as the index of expected agreement, which is the proportion of cases in which the results of both assessments matched.33

RESULTS

Response rates and study population

The country coordinators approached a total of 1396 hospitals, of which 483 visited the web-based questionnaire and ultimately 389 submitted the completed questionnaire (table 1). Response rates varied per country. The study population consisted of public (80%) as well as private (20%) hospitals, and included university (23.5%), teaching (48.9%) and non-teaching hospitals (27.6%). The mean number of sites per hospital was 2.46. Almost a quarter of all hospitals were considering collaboration to deliver cross-border patient care, but few hospitals were doing so at the time of the study.15
Table 1

Hospital recruitment and response rates by country

Hospitals approachedHospitals entering web-based questionnaireHospitals concluding web-based questionnaire
UK2504114
Ireland442925
The Netherlands971210
Belgium453325
France32210078
Spain307131113
Poland2508480
Czech Republic815344
Total1396483389

Results of the validation process

Regarding reliability, comparison of the information obtained for 25 items from questionnaires and on-site visits resulted in the following ranges of agreement: for five items agreement was over 90%, for 12 items the level of agreement was >70% and <90%, and for eight items agreement was <70%. Given the period of 7–9 months between the questionnaire and the on-site visits, it is likely that at least some of the items studied had actually changed. In addition, the 14 items analysed to assess criterion validity resulted in the following levels of agreement: seven items had an agreement index of >90%, six items rated >70% and <90%, and one item scored an index of <70%. Based on these results, we considered the information collected with the questionnaire a fair approximation of the actual situation at participating hospitals.

QI strategies

A previous MARQuIS study34 identified seven mandatory QI strategies for hospitals. The extent to which these strategies are applied at European hospitals is reported below.

QI strategy 1: Organisational quality management programmes

Hospitals reported using quality management programmes in developing and implementing QI (table 2). Overall, the International Organization for Standardization (ISO) 9000 management system standards were used most often, and the European Foundation for Quality Management (EFQM) model was used least often. However, there were wide variations between countries. Belgium was the only country where the EFQM model was preferred by most hospitals (60.9%); in all other countries the ISO system was the dominant scheme. In Poland (50.0%), the Czech Republic (59.5%), and Spain (70.3%) the use of ISO was widespread. In Spain some hospitals seemed to use both schemes. Irish hospitals reported moderate use of a quality management programme (ISO = 31.6%, EFQM 20.0%), but stressed the systematic use of specific QI teams in most of their hospital departments, either systematically (47.8%) or unsystematically (17.4%). French hospitals relied least on the programmatic approach with the ISO (28.6%) or EFQM system (3.4%), and the availability if QI teams was limited (27.7%). Responsibilities for, and policies on, blood transfusion, the use of antibiotics, and hospital infection control were clearly assigned to a committee or person in almost all hospitals across Europe, with the exception of Czech hospitals, which reported significantly less clear organisation for all three hospital-wide functions. Spanish hospitals showed a gap in organising responsibilities for blood transfusion (87.5%), and Irish hospitals in the use of antibiotics (87%). Responsibility for the prevention of bed sores seemed less structured within European hospitals, varying from a reported 66.7% in Ireland to 95.8% in Belgium.

QI strategy 2: Systems for obtaining patients’ views

Monitoring the views of patients by systematically conducting patient surveys was a common practice in 64.5% of the participating European hospitals. The Czech Republic stood out, reporting that 91.9% of their hospitals systematically monitored patient views. These numbers refer to hospital-wide systems for collecting patients’ views on the care they received. At the department level, patients are asked at discharge for their opinion on the quality of care delivered by the hospital staff. In France this strategy was widely implemented, with approximately 65% of the hospitals reporting a policy to measure patients’ opinion at discharge. In Poland, this was a common practice in less than 14% of the hospitals. The rates varied significantly for the three medical conditions included in our study, and variation between countries seemed greater than within countries (table 3).
Table 3

Quality improvement (QI) strategies as applied in European hospitals: systems for getting the views of patients (total and per country,* numbers are valid percentages†)

Specification of QI strategyTotalIrelandBelgiumFranceSpainPolandCzech Republic
Monitoring the views of patients/performing patient surveys (n = 344)3442223661047537
    164.554.560.956.167.361.391.9
    21418.213.013.611.514.78.1
    315.727.313.021.216.316.00
    44.708.77.64.85.30
Analysis of patient complaints (n = 350)3502324661067338
    185.782.683.383.388.784.092.1
    26.64.38.36.12.88.07.9
    35.48.74.29.15.75.30
    41.74.34.21.52.800
Patient involvement in‡:
    Discussing the results of patient surveys and complaints handling (n = 345)3452324631057538
        17.24.3020.61.94.07.9
        26.74.38.319.01.04.02.6
        319.130.412.530.217.19.315.8
        460.652.270.822.274.376.068.4
    The development of quality criteria/standards (n = 348)3482324651067538
        11.7001.501.37.9
        22.38.702.304.00
        316.134.816.716.97.59.310.5
        473.652.270.872.386.877.376.3
    Designing protocols (n = 345)3452224651047538
        11.200002.75.3
        21.74.503.102.70
        315.127.38.324.68.76.70
        476.563.683.370.886.577.389.5
    The evaluation of achieving quality objectives (n = 347)3472324661047538
        14.908.36.11.08.05.3
        27.213.08.39.11.910.72.6
        320.556.516.740.98.78.07.9
        461.426.158.342.482.761.378.9
    Participation in (quality) committees (n = 345)3452324661047337
        17.08.7025.81.01.42.7
        22.68.706.1002.7
        318.630.412.547.012.52.70
        465.543.579.221.280.882.289.2
    Participation in improvement projects (n = 342)3422222661037438
        13.59.106.11.01.47.9
        25.89.1010.61.96.80
        327.540.922.753.021.413.55.3
        457.036.468.230.367.068.981.6
Patients’ opinion about quality of care asked at discharge (for patients with acute myocardial infarction) (n = 319)3192224491026937
    134.522.737.569.439.28.740.5
    215.413.612.512.212.715.927.0
    323.818.237.58.222.524.629.7
    420.731.88.36.122.543.52.7
Patients’ opinion about quality of care asked at discharge (for patients with appendicitis) (n = 313)3132221511007134
    139.622.747.676.544.08.547.1
    214.718.24.813.710.019.717.6
    319.818.223.83.924.021.120.6
    419.222.714.32.021.036.611.8
Patients’ opinion about quality of care asked at discharge (patients at maternal service) (n = 301)301182450877334
    144.927.841.750.044.824.750.0
    214.611.116.711.814.920.511.8
    314.011.116.723.514.915.123.5
    415.322.220.82.917.224.72.9

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates. †The sum of percentages may not always equal 100%; the percentage answers “don’t know/no answer” are not listed here. 1 =  yes, in most departments (>50%); 2 =  yes, in most departments (>50%), but not systematically; 3 =  yes, in some departments (<50%); 4 =  no. ‡From this point on the answer categories 1–4 should be read as: 1 =  yes always; 2 =  most of the time; 3 =  sometimes; 4 =  no.

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates. †The sum of percentages may not always equal 100%; the percentage answers “don’t know/no answer” are not listed here. 1 =  yes, in most departments (>50%); 2 =  yes, in most departments (>50%), but not systematically; 3 =  yes, in some departments (<50%); 4 =  no. ‡From this point on the answer categories 1–4 should be read as: 1 =  yes always; 2 =  most of the time; 3 =  sometimes; 4 =  no. Across participating European hospitals, patient involvement seemed to be little developed. Hospitals were asked to identify the activities in which individual patients or patient organisations were always or almost always involved. Participation in the design of protocols or the development of standards was reported by 3% to 4% of all hospitals; participation in improvement projects or in quality committees was reported by less than 10%. Patient involvement was best implemented in France, with almost 40% of the hospitals reporting that they involved patients in the discussion of the results of patient surveys or complaints, and 32% stating that patients participated in quality committees.

QI strategy 3: Patient safety systems

Hospitals were asked how patient safety was organised and managed, whether the results were reported, and if so, how they were reported (table 4). Responsibility for patient safety was assigned to a committee or person in approximately 75% of the hospitals. At 39.1% of the hospitals a risk management programme or system was in place; 50% of the hospitals systematically reported and analysed adverse events, and 55.6% also reported complications to the medical staff. These are average numbers for Europe; variation between countries was substantial. Irish hospitals scored consistently high (>90%), and Belgium and Spanish hospitals relatively low on the availability of safety systems.
Table 4

Quality improvement (QI) strategies as applied in European hospitals: specific patient safety systems; total and per country*, numbers are valid percentages (total item response in absolute numbers)

Specification of QI strategyTotalIrelandBelgiumFranceSpainPolandCzech Republic
Adverse event reporting and analysis† (n = 348)3482224661067537
    150.090.920.856.130.250.783.8
    214.44.58.318.213.214.710.8
    328.24.558.322.743.425.35.4
    46.008.33.012.35.30
Risk management programme/system† (n = 348)3482324661067338
    139.191.316.740.934.921.957.9
    213.84.34.216.79.415.123.7
    325.34.354.239.430.215.15.3
    418.7016.73.023.638.413.2
Patient safety person/groupy = 73.5 (347)y = 95.7 (23)y = 58.3 (24)y = 83.1 (65)y = 75.0 (104)y = 54.1 (74)y = 74.4 (39)
Patient identification systems: use of bracelets in the emergency departmenty = 39.8 (334)y = 56.5 (23)y = 63.6 (22)y = 37.3 (61)y = 38.8 (103)y = 36.4 (77)y = 14.7 (34)
Patient identification systems: use of bracelets for admitted patientsy = 46.5 (333)y = 100 (23)y = 90.9 (22)y = 29.5 (61)y = 42.2 (102)y = 35.1 (77)y = 29.4 (34)
Drug safety:
    Standardised limited number of drugsy = 91.4 (348)y = 73.9 (23)y = 95.8 (24)y = 84.6 (65)y = 99.0 (105)y = 98.7 (76)y = 73.7 (38)
    Electronic drug prescription systemy = 39.4 (350)y = 13.0 (23)y = 25.0 (24)y = 33.3 (66)y = 38.7 (106)y = 38.2 (76)y = 86.8 (38)
    Expiration date checked (AMI)y = 91.4 (315)y = 81.8 (22)y = 91.7 (24)y = 84.8 (46)y = 90.1 (101)y = 97.1 (70)y = 94.7 (38)
    Drugs locked (AMI)y = 73.5 (317)y = 81.8 (22)y = 75.0 (24)y = 69.6 (46)y = 43.1 (102)y = 98.6 (70)y = 97.4 (38)
    HR drugs separately stored (AMI)y = 89.9 (316)y = 85.7 (21)y = 91.7 (24)y = 80.4 (46)y = 90.2 (102)y = 97.1 (70)y = 84.2 (38)
MRSA testing (AMI)y = 6.9 (306)y = 4.8 (21)y = 4.2 (24)y = 4.7 (43)y = 0 (99)y = 14.9 (67)y = 5.4 (37)
Reporting complications to medical staffy = 55.6 (304)y = 44.4 (18)y = 35.0 (20)y = 22.8 (57)y = 63.6 (88)y = 58.3 (72)y = 85.7 (35)

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates.

†The sum of percentages may not always equal 100%; the percentage answers ‘don’t know/no answer’ are not listed here. 1 =  yes, in most departments (>50%); 2 =  yes, in most departments (>50%), but not systematically; 3 =  yes, in some departments (<50%); 4 =  no. y, yes.

AMI, acute myocardial infarction; MRSA, methicillin resistant Staphylococcus aureus.

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates. †The sum of percentages may not always equal 100%; the percentage answers ‘don’t know/no answer’ are not listed here. 1 =  yes, in most departments (>50%); 2 =  yes, in most departments (>50%), but not systematically; 3 =  yes, in some departments (<50%); 4 =  no. y, yes. AMI, acute myocardial infarction; MRSA, methicillin resistant Staphylococcus aureus. Specific safety questions addressed drug safety management and patient identification. In general, drug safety seemed to be assured in all participating hospitals: the use of drugs was standardised and controlled, and systems for storing, checking and preventing unauthorised access to drugs seemed well implemented. Electronic drug prescriptions were used widely only in Czech hospitals (86.8%). By comparison, in Ireland a mere 13% of the hospitals reported use of electronic prescriptions. For patient identification systems the findings were the opposite, with 100% of Irish versus 29.4% of Czech hospitals using bracelets to identify admitted patients.

QI strategy 4: Clinical guidelines

Clinical guidelines were widely used at participating European hospitals. Hospital-wide guidelines for preoperative assessment and prophylactic antibiotic use were in place in the vast majority (75–90%) of hospitals. In the Czech Republic and Ireland, guidelines for prophylactic antibiotic use were least widely used, versus 100% coverage in Belgian hospitals. Laboratory work seemed to be highly standardised across the various types of laboratories across Europe. On average, standard operating procedures (SOPs) were available in approximately 90% of all hospitals. At the department level, availability of clinical guidelines or protocols was significantly less common. In summary, guidelines were available for the management of patients with AMI (mean  = 86.7%), appendicitis (mean  = 54.3%), or obstetrical problems such as breech presentation (71.5%) and vaginal birth after caesarean delivery (64.5%). Between-country variation for the availability of clinical or practice guidelines was limited for hospital-wide guidelines and SOPs, but substantial for condition-related, department-level guidelines (table 5).
Table 5

Quality improvement (QI) strategies as applied in European hospitals: clinical and practice guidelines; total and per country*, numbers are positive responses in valid percentages (total item response in absolute numbers)

Specification of QI strategyTotalIrelandBelgiumFranceSpainPolandCzech Republic
Hospital-wide guidelines:
    Preoperative assessment74.8 (306)72.7 (22)90.5 (21)57.4 (47)71.4 (98)75.7 (70)88.2 (34)
    Use of antibiotics83.0 (311)63.6 (22)100 (22)88.2 (51)82.5 (97)92.9 (70)51.4 (35)
    Prophylactic use of antibiotics89.6 (309)66.7 (21)100 (22)94.1 (51)94.8 (97)91.4 (70)71.4 (35)
Standard operating procedures for various types of laboratories:
    Clinical chemistry93.5 (310)95.5 (22)100 (22)88.5 (52)94.7 (95)93.8 (65)89.5 (38)
    Pathology72.5 (265)95.7 (23)76.2 (21)71.0 (31)82.6 (92)46.0 (50)58.8 (34)
    Microbiology laboratory87.6 (298)95.2 (21)100 (22)89.8 (49)90.5 (95)85.9 (64)66.7 (33)
    Pharmacy86.9 (314)91.3 (23)86.4 (22)90.2 (61)92.7 (96)87.7 (65)59.4 (32)
    Diagnostic radiology89.0 (310)86.4 (22)100 (21)83.9 (56)92.6 (95)89.7 (68)79.4 (34)
Clinical guidelines for AMI: management of AMI patients86.7 (316)72.7 (22)95.8 (24)89.1 (46)89.3 (103)79.4 (68)86.8 (38)
Clinical guidelines for appendicitis:
    Management of suspected appendicitis54.3 (311)36.4 (22)47.8 (23)52.0 (50)49.5 (97)67.6 (71)51.4 (35)
    Wrong site, wrong surgery42.0 (307)68.2 (22)54.5 (22)47.9 (48)15.6 (96)57.7 (71)29.4 (34)
Clinical guidelines for obstetrics:
    Breech presentation71.5 (281)58.8 (17)81.8 (22)61.7 (47)60.8 (79)90.0 (70)59.4 (32)
    VBAC64.5 (293)35.3 (17)52.0 (25)64.6 (48)59.0 (83)83.3 (72)50.0 (34)

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates.

AMI, acute myocardial infarction; VBAC, vaginal birth after caesarean section.

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates. AMI, acute myocardial infarction; VBAC, vaginal birth after caesarean section.

QI strategy 5: Performance indicators or measures

For the three medical conditions included in this study, hospitals were asked to report the availability of performance data for a selection of clinical indicators. Table 6 shows the findings. In summary, the availability of AMI performance data was most complete, averaging approximately 70% for the seven selected indicators. Poland reported the highest percentages, France the lowest. Performance data on the management of appendicitis were being collected for approximately 50% of the five indicators, varying from 42% for perforated appendicitis treated surgically 24 h after admission to 68.2% for wound infections. The Czech Republic and Poland performed best in this area. Lastly, the statistics for obstetrical indicators varied from 54.0% for the rate of vaginal birth after caesarean delivery to 85.3% for the percentage of caesarian deliveries. Obstetrical data were most complete in Poland and Belgium, and least complete in Ireland (table 6).
Table 6

Quality improvement (QI) strategies as applied in European hospitals: performance indicators or measures; total and per country*, numbers are positive responses in valid percentages (total item response in absolute numbers)

Specification of QI strategyTotalIrelandBelgiumFranceSpainPolandCzech Republic
Availability of AMI performance indicators:
    Door-to-needle time57.7 (291)68.2 (22)62.5 (24)45.2 (42)62.0 (92)51.6 (64)55.9 (34)
    Receipt of reperfusion70.3 (293)54.5 (22)75.0 (24)61.4 (44)66.7 (93)79.4 (63)76.5 (34)
    Aspirin use <24 h71.8 (291)68.2 (22)79.2 (24)59.1 (44)69.2 (91)82.5 (63)70.6 (34)
    Prescription ACE inhibitors at discharge67.5 (292)63.6 (22)73.9 (23)53.5 (43)59.1 (93)84.4 (64)67.6 (34)
    Prescription of β-blockers at discharge71.2 (288)63.6 (22)73.9 (23)51.2 (43)69.6 (92)87.3 (63)71.9 (32)
    Prescription of aspirin at discharge73.6 (288)68.2 (22)83.3 (24)58.1 (43)71.4 (91)85.5 (62)69.7 (33)
    Inpatient mortality74.4 (289)57.1 (21)79.2 (24)54.8 (42)87.0 (92)77.4 (62)70.6 (34)
Availability of performance indicators for the management of appendicitis:
    Prophylactic antibiotics53.8 (290)33.3 (21)43.5 (23)43.2 (44)65.2 (92)73.5 (68)13.3 (30)
    Negative appendectomy46.5 (288)33.3 (21)43.5 (23)42.2 (45)45.1 (91)53.0 (66)50.0 (30)
    Rate of lap versus open appendectomy51.0 (286)47.6 (21)60.9 (23)62.2 (45)58.7 (92)21.5 (65)64.3 (28)
    Perforated appendicitis operated 24 h after admittance42.0 (283)20.0 (20)39.1 (23)29.5 (44)40.7 (91)50.8 (63)56.7 (30)
    Wound infections68.2 (280)50.0 (20)47.8 (23)47.7 (44)77.8 (90)75.4 (61)80.0 (30)
Availability of performance indicators for deliveries:
    Induced labour rate67.5 (280)58.8 (17)75.0 (20)61.7 (47)60.5 (81)76.8 (69)62.5 (32)
    % Caesarean sections of total deliveries85.3 (278)64.7 (17)90.0 (20)83.7 (49)87.7 (81)91.0 (67)74.2 (31)
    VBAC rate54.0 (274)35.3 (17)50.0 (20)44.7 (47)55.8 (77)67.6 (68)51.6 (31)
    Deliveries with peridural anaesthesia71.9 (278)52.9 (17)85.0 (20)83.3 (48)81.3 (80)58.8 (68)54.8 (31)

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates.

AMI, acute myocardial infarction; VBAC, vaginal birth after caesarean section.

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates. AMI, acute myocardial infarction; VBAC, vaginal birth after caesarean section.

QI strategy 6: Internal audit, assessment of clinical standards

Medical staff performance was systematically reviewed at 50% of the participating hospitals, and peer review (site visits) was conducted at approximately 25%. Between-country variations were considerable. Belgium, Poland and the Czech Republic reported that over 60% of the hospitals performed medical staff performance reviews, versus 26.1% of Irish hospitals. However, Irish hospitals made more use of peer review (site visits) than any other European country (39.1%). On average, 50% of the laboratories at European hospitals were periodically surveyed by an internal audit team. Percentages varied according to the type of laboratory and between countries. France reported generally low rates; in Poland internal auditing seemed broadly implemented. However, only a third of the Polish hospitals reported the results of internal audits to their governing boards, versus approximately 90% of the hospitals in the Czech Republic and Ireland. Polish hospitals more openly shared the results with their medical staffs (59.2%), but other countries reported higher percentages. Belgium was the exception here: only 40% disclosed their results to medical staff (table 7).
Table 7

Quality improvement (QI) strategies as applied in European hospitals: audit, internal assessment of clinical standards; total and per country*, numbers are valid percentages (total item response in absolute numbers)

Specification of QI strategyTotalIrelandBelgiumFranceSpainPolandCzech Republic
Internal auditing of hospital departments† (n = 347)3472224651057538
    135.722.716.716.921.953.376.3
    211.013.612.518.511.44.05.3
    340.163.654.256.952.416.013.2
    410.1012.57.714.314.72.6
Peer review/visitatie of hospital departments† (n = 341)3412324621047636
    117.326.120.88.17.721.138.9
    28.513.016.73.28.76.68.3
    329.943.533.345.227.914.522.2
    437.513.020.841.951.042.125.0
Medical staff performance review† (n = 345)3452324651037538
    136.717.441.738.523.352.042.1
    214.58.720.812.311.716.021.1
    323.752.225.015.432.010.723.7
    421.121.78.326.230.114.713.2
Periodical internal audits of laboratories:
    Clinical chemistry laboratoryy = 60.3 (300)y = 68.2 (22)y = 77.3 (22)y = 20.0 (33)y = 46.2 (93)y = 78.8 (66)y = 81.1 (37)
    Pathology laboratoryy = 37.4 (254)y = 72.7 (22)y = 28.6 (21)y = 3.7 (27)y = 34.1 (88)y = 30.6 (49)y = 44.1 (34)
    Microbiology laboratoryy = 54.3 (278)y = 81.0 (21)y = 81.8 (22)y = 17.1 (41)y = 42.4 (92)y = 74.1 (58)y = 51.6 (31)
    Pharmacy laboratoryy = 47.5 (297)y = 54.5 (22)y = 18.2 (22)y = 21.8 (55)y = 45.1 (91)y = 73.8 (61)y = 50.0 (32)
    Diagnostic radiology laboratoryy = 48.5 (291)y = 60.0 (20)y = 25.0 (20)y = 12.2 (49)y = 41.1 (90)y = 72.6 (62)y = 66.7 (36)
Results of internal audits are formally reported to:
    Hospital’s governing boardy = 65.0 (343)y = 91.3 (23)y = 62.5 (24)y = 44.4 (63)y = 79.6 (103)y = 34.2 (73)y = 89.7 (39)
    Medical staffy = 67.4 (304)y = 73.7 (19)y = 40.0 (20)y = 72.9 (59)y = 65.1 (83)y = 59.2 (71)y = 91.4 (35)

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates. †The sum of percentages may not always equal 100%; the percentage answers “don’t know/no answer” are not listed here. 1 =  yes, systematically in most departments (>50%); 2 =  yes, in most departments (>50%), but not systematically; 3 =  yes, in some departments (<50%); 4 =  no.

y, yes.

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates. †The sum of percentages may not always equal 100%; the percentage answers “don’t know/no answer” are not listed here. 1 =  yes, systematically in most departments (>50%); 2 =  yes, in most departments (>50%), but not systematically; 3 =  yes, in some departments (<50%); 4 =  no. y, yes.

QI strategy 7: External assessment

Most hospitals (88%) have been assessed (at least in part) by an external organisation such as an accreditation (59.4%) or certification (49.4%) institute, a patient organisation (18.5%), or a government inspection body (66%). Some hospitals were audited by more than one organisation. In Spain, for instance, 64.8% of all hospitals (n = 88) reported being evaluated by an accreditation body, and 63.6% by a certification institute. In France (n = 63), 93.7% of all hospitals had been accredited, and in Ireland (n = 22) 90.9%. In Poland (n = 75), government inspections were the most frequently reported type of external evaluation (76%). French hospitals reported being most open (92.3%), and Spanish hospitals the least open (19.8%) about their assessment results. On average, 52.9% of the hospitals in our sample publicly disclosed their assessment results (table 8). Most participating hospitals (84.3%) reported plans for re-evaluation within the next 3 years (not shown). Accreditation bodies were listed most frequently as the future assessors; in the Czech Republic, Ireland, the UK, and the Netherlands, more than 85% of the hospitals expressed this intention, and the figures were 78.1% for France and 77.3% for Spain.
Table 8

Quality improvement (QI) strategies as applied in European hospitals: external assessment, schemes and programmes; total and per country*, numbers are positive responses in valid percentages (total item response in absolute numbers)

Specification of QI strategyTotalIrelandBelgiumFranceSpainPolandCzech Republic
(Part of) hospital previously externally assessed88.0 (351)100 (23)95.8 (24)96.9 (65)85.8 (106)82.9 (76)71.8 (39)
External assessment by:
    Accreditation institute59.4 (323)90.9 (22)20.8 (24)93.7 (63)64.8 (88)35.5 (76)29.4 (34)
    Certification institute49.4 (314)60.0 (20)41.7 (24)23.3 (60)63.6 (88)48.6 (72)48.6 (37)
    Patient/consumer organisation18.5 (297)33.3 (18)34.8 (23)3.4 (58)5.3 (76)18.9 (74)36.4 (33)
    Inspection66.0 (318)80.0 (20)83.3 (24)55.9 (59)65.9 (91)76.0 (75)28.1 (32)
Public disclosure of assessment results52.9 (331)60.9 (23)33.3 (24)92.3 (65)19.8 (91)50.6 (77)67.6 (34)
Previous external assessment of laboratories:
    Clinical chemistry laboratory68.6 (303)57.1 (21)76.2 (21)65.2 (46)59.4 (96)70.8 (65)81.6 (38)
    Pathology laboratory38.7 (261)60.9 (23)19.0 (21)50.0 (28)34.8 (92)28.6 (49)35.3 (34)
    Microbiology laboratory58.7 (283)65.0 (20)71.4 (21)60.5 (43)48.4 (93)68.3 (60)50.0 (32)
    Pharmacy laboratory54.6 (302)72.7 (22)59.1 (22)65.5 (58)39.6 (91)62.9 (62)39.4 (33)
    Diagnostic radiology laboratory56.8 (294)85.0 (20)60.0 (20)61.2 (49)47.3 (93)56.5 (62)60.0 (35)

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates.

*The results for the UK and the Netherlands are included in the total but not listed separately, due to the very low response rates. Quality improvement (QI) strategies are widely used in European hospitals. The most widely applied QI strategy is external assessment of hospitals, whereas patient involvement in QI activities is the least widely applied Reported implementation varies per country. This leaves considerable room for progress in making QI in hospitals a reality Differences also suggest that, for various reasons, countries may prefer some QI strategies over others International comparisons of the use of QI strategies can promote learning and the spread of good practice The results of this study may be useful to national policy makers in monitoring the attainment of healthcare policy goals

Key messages

Points for further research

Further research should focus on exploring the relationship between the use of QI strategies and the actual performance of hospitals, including the relative contribution of each of the seven QI strategies to performance

DISCUSSION

This study has some limitations. The response rates varied per country, and were particularly low in the UK and the Netherlands. This may be explained by the various approaches used to recruit hospitals for the MARQuIS project, and to the effect of questionnaire fatigue due to the over-application of questionnaire surveys to evaluate healthcare performance in general. We therefore cannot rule out participation bias. Also, accuracy of the information is always a limitation when using self-reported data. However, the results of our validation process strongly suggest that the reported results are fairly accurate. Further, translation of the questionnaire, the use of jargon, and the involvement of people from various healthcare systems may have caused differences in how the items were interpreted. Lastly, hospitals may use local QI approaches or tools not included in this questionnaire, in which case the application of QI strategies, as described in this article, may misrepresent the “maturity” of hospitals’ quality management systems. These limitations should be taken into account when interpreting results. International comparisons can promote learning and the spread of good practice, and are one of the ways in which the European Community is expected to raise healthcare quality. This study of how European hospitals apply seven common QI strategies found considerable variation between the level of implementation of the different strategies—a finding that leaves considerable scope for progress in making QI a reality. The use of QI strategies at the European level was determined or at least influenced by national and international policy making and regulation, as well as by national and local bottom-up actions initiated by professionals or others.35 In our study 88% of all hospitals reported having been externally assessed; the widespread application of the “external assessment” QI strategy can be ascribed to the fact that most countries have adopted one or more models of external assessment (ie, accreditation, certification or licensure) to ensure and improve hospital performance, which in turn has been related to financing healthcare delivery. However, policies and regulations may not always be effective, as shown by the fact that in most hospitals (>90%), patient involvement in QI activities was lacking. This was despite the various legal and other efforts undertaken by the European Commission over the past decades to increase citizens’ participation in QI, and in the organisation and structure of health services in general.4 5 Future research should focus on detecting barriers to the implementation of these QI strategies. In this regard, efforts by the EU to facilitate improvements and foster European collaboration may help to further increase implementation.36 37 Legislation recently proposed by the European Commission stresses the values and principles of safe, high-quality health services that underpin European health systems. However, the question arises as to how these agreed-upon values and principles can be applied by member states.36 We believe our results may help national policy makers to monitor the attainment of healthcare policy goals. The application of more QI strategies, however, may not necessarily imply more positive effects on performance. Our findings would be even more valuable if the demonstrated use of QI strategies could be related to actual performance in hospitals. This would give EU policy makers direct input for monitoring the development of healthcare policies and regulations. Elsewhere in this supplement this relationship is explored in greater depth.
  24 in total

1.  Development of a healthcare quality improvement measurement tool: results of a content validity study.

Authors:  Steven J Meurer; Doris McGartland Rubio; Michael A Counte; Tom Burroughs
Journal:  Hosp Top       Date:  2002

2.  Relationship of hospital teaching status with quality of care and mortality for Medicare patients with acute MI.

Authors:  J J Allison; C I Kiefe; N W Weissman; S D Person; M Rousculp; J G Canto; S Bae; O D Williams; R Farmer; R M Centor
Journal:  JAMA       Date:  2000-09-13       Impact factor: 56.272

3.  Quality improvement initiative and its impact on the management of patients with acute myocardial infarction.

Authors:  R H Mehta; S Das; T T Tsai; E Nolan; G Kearly; K A Eagle
Journal:  Arch Intern Med       Date:  2000-11-13

4.  Value of abdominal CT in the emergency department for patients with abdominal pain.

Authors:  Max P Rosen; Bettina Siewert; Daniel Z Sands; Rebecca Bromberg; Jonathan Edlow; Vassilios Raptopoulos
Journal:  Eur Radiol       Date:  2002-10-16       Impact factor: 5.315

5.  Does the ownership of the admitting hospital make a difference? Outcomes and process of care of Medicare beneficiaries admitted with acute myocardial infarction.

Authors:  Frank A Sloan; Justin G Trogdon; Lesley H Curtis; Kevin A Schulman
Journal:  Med Care       Date:  2003-10       Impact factor: 2.983

6.  Laparoscopic versus open appendectomy: outcomes comparison based on a large administrative database.

Authors:  Ulrich Guller; Sheleika Hervey; Harriett Purves; Lawrence H Muhlbaier; Eric D Peterson; Steve Eubanks; Ricardo Pietrobon
Journal:  Ann Surg       Date:  2004-01       Impact factor: 12.969

7.  Enhancing quality of care for acute myocardial infarction: shifting the focus of improvement from key indicators to process of care and tool use: the American College of Cardiology Acute Myocardial Infarction Guidelines Applied in Practice Project in Michigan: Flint and Saginaw Expansion.

Authors:  Rajendra H Mehta; Cecelia K Montoye; Jessica Faul; Dorothy J Nagle; James Kure; Ethiraj Raj; Peter Fattal; Shiraz Sharrif; Mohamadali Amlani; Hameem U Changezi; Stephen Skorcz; Nancy Bailey; Theresa Bourque; Mary LaTarte; Donna McLean; Suzanne Savoy; Paul Werner; Patricia L Baker; Anthony DeFranco; Kim A Eagle
Journal:  J Am Coll Cardiol       Date:  2004-06-16       Impact factor: 24.094

8.  Birthing positions. A qualitative study into the views of women about various birthing positions.

Authors:  A De Jonge; A L M Lagro-Janssen
Journal:  J Psychosom Obstet Gynaecol       Date:  2004-03       Impact factor: 2.949

Review 9.  Systematic review: computed tomography and ultrasonography to detect acute appendicitis in adults and adolescents.

Authors:  Teruhiko Terasawa; C Craig Blackmore; Stephen Bent; R Jeffrey Kohlwes
Journal:  Ann Intern Med       Date:  2004-10-05       Impact factor: 25.391

10.  Do European hospitals have quality and safety governance systems and structures in place?

Authors:  C Shaw; B Kutryba; H Crisp; P Vallejo; R Suñol
Journal:  Qual Saf Health Care       Date:  2009-02
View more
  31 in total

1.  Clinical audit: a useful tool for reducing severe postpartum haemorrhages?

Authors:  Corinne Dupont; Catherine Deneux-Tharaux; Sandrine Touzet; Cyrille Colin; Marie-Hélène Bouvier-Colle; Jacques Lansac; Simone Thevenet; Claire Boberie-Moyrand; Gaëlle Piccin; Marie-Pierre Fernandez; René-Charles Rudigoz
Journal:  Int J Qual Health Care       Date:  2011-07-06       Impact factor: 2.038

Review 2.  [Patient safety in anesthesiology and intensive care medicine. Measures for improvement].

Authors:  C Rosenthal; F Balzer; W Boemke; C Spies
Journal:  Med Klin Intensivmed Notfmed       Date:  2012-11-07       Impact factor: 0.840

Review 3.  A systematic review and meta-synthesis of policy intervention characteristics that influence the implementation of government-directed policy in the hospital setting: implications for infection prevention and control.

Authors:  Sally M Havers; Elizabeth Kate Martin; Andrew Wilson; Lisa Hall
Journal:  J Infect Prev       Date:  2020-05-04

4.  Measuring patient experience: concepts and methods.

Authors:  Faraz Ahmed; Jenni Burt; Martin Roland
Journal:  Patient       Date:  2014       Impact factor: 3.883

5.  Contribution of Healthcare Professionals in Issues that Relate to Quality Management.

Authors:  Mary Kyriacou Georgiou; Anastasios Merkouris; Maria Hadjibalassi; Pavlos Sarafis
Journal:  Mater Sociomed       Date:  2021-03

6.  Understanding and benchmarking health service achievement of policy goals for chronic disease.

Authors:  Erica Bell; Bastian Seidel
Journal:  BMC Health Serv Res       Date:  2012-09-29       Impact factor: 2.655

7.  Implementation of patient safety strategies in European hospitals.

Authors:  R Suñol; P Vallejo; O Groene; G Escaramis; A Thompson; B Kutryba; P Garel
Journal:  Qual Saf Health Care       Date:  2009-02

8.  Cross-border care and healthcare quality improvement in Europe: the MARQuIS research project.

Authors:  R Suñol; P Garel; A Jacquerye
Journal:  Qual Saf Health Care       Date:  2009-02

9.  Differentiating between hospitals according to the "maturity" of quality improvement systems: a new classification scheme in a sample of European hospitals.

Authors:  M J M H Lombarts; I Rupp; P Vallejo; N S Klazinga; R Suñol
Journal:  Qual Saf Health Care       Date:  2009-02

10.  Evaluating a questionnaire to measure improvement initiatives in Swedish healthcare.

Authors:  Ann-Christine Andersson; Mattias Elg; Kent-Inge Perseius; Ewa Idvall
Journal:  BMC Health Serv Res       Date:  2013-02-07       Impact factor: 2.655

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.