Literature DB >> 23015098

Method for developing national quality indicators based on manual data extraction from medical records.

Melanie Couralet1, Henri Leleu, Frederic Capuano, Leah Marcotte, Gérard Nitenberg, Claude Sicotte, Etienne Minvielle.   

Abstract

Developing quality indicators (QI) for national purposes (eg, public disclosure, paying-for-performance) highlights the need to find accessible and reliable data sources for collecting standardised data. The most accurate and reliable data source for collecting clinical and organisational information still remains the medical record. Data collection from electronic medical records (EMR) would be far less burdensome than from paper medical records (PMR). However, the development of EMRs is costly and has suffered from low rates of adoption and barriers of usability even in developed countries. Currently, methods for producing national QIs based on the medical record rely on manual extraction from PMRs. We propose and illustrate such a method. These QIs display feasibility, reliability and discriminative power, and can be used to compare hospitals. They have been implemented nationwide in France since 2006. The method used to develop these QIs could be adapted for use in large-scale programmes of hospital regulation in other, including developing, countries.

Entities:  

Mesh:

Year:  2012        PMID: 23015098      PMCID: PMC3582043          DOI: 10.1136/bmjqs-2012-001170

Source DB:  PubMed          Journal:  BMJ Qual Saf        ISSN: 2044-5415            Impact factor:   7.035


Introduction

Over the past 10 years, the use of quality indicators (QIs) has been strongly encouraged as a means of assessing quality in hospitals. QIs are now a widely used tool in hospital regulation in developed countries (eg, in performance-based financing and the public disclosure of hospital comparisons).1 It is well established that key attributes of QIs are able to detect targeted areas for improvement on topics of importance, scientific soundness and feasibility.2 For national comparisons of healthcare organisations (HCO), a valid and standardised data collection process is also required, as any errors could affect a hospital's reputation and also have financial repercussions. Three main data sources are used to develop QIs: (1) ad hoc surveys (eg, patient's experience and satisfaction indicators), but these are costly and require recruitment of respondents and high hospital commitment,3 (2) medico-administrative data (eg, patient safety indicators), but these often capture limited information on complex care processes4 and (3) medical records (eg, clinical practice and organisational indicators) which are the preferred option for obtaining accurate and reliable clinical and organisational information.5–9 Most medical records are still paper medical records (PMRs) and entail difficulties in terms of data extraction that remains manual.10 The adoption of interoperable electronic medical record (EMR) systems could promote efficiency by developing an automated process of data extraction. However, it is expensive. Moreover, in the most highly developed countries that invest in this area, it remains arduous. For instance in US hospitals, only 13% reported use of a basic EMR system in 2008, according to a study by Jha et al.11 Although these numbers have significantly increased over the last few years—2011 data shows 35% adoption of basic EMR systems by US hospitals—the rates of adoption are still low.12 A recent national study shows that in France only 6% of medical records are fully electronic.13 In terms of data extraction for the purposes of quality measurement, a basic EMR system does not necessarily enable easy and automatic computation of aggregated data, nor does it preclude the use of partial paper charts, making some data completely inaccessible via the EMR. We propose a pragmatic method for using PMRs to produce national QIs that display feasibility, reliability and discriminative power, and that enable PMR audits for hospital comparison. The method is based on data extraction from a random sample of PMRs in each hospital. It has been implemented in France since 2006, but could be adopted by other countries interested in assessing large-scale hospital performance.14 We describe the methods development, the PMR sampling strategy and the statistical procedures for ensuring robustness. Each step of the method is illustrated with appropriate examples. Last, we discuss the place of such a method in the context of development of EMR systems.

Background to method development and implementation

The method was developed and has been implemented by the COMPAQH project team (COordination for Measuring Performance and Assuring Quality in Hospitals), a French national initiative for the development and use of QIs, coordinated by the French Institute for Health and Medical Research (INSERM) and sponsored by the Ministry of Health and the French National Authority for Health (Haute Autorité de Santé, HAS).13 15 The QIs were selected and designed with the collaboration of the French public authorities and health professionals. Each is categorised within nine priority areas for quality improvement: (1) pain management; (2) continuity of care; (3) management of patients with nutritional disorders; (4) patient safety; (5) taking account of patients’ views; (6) implementation of practice guidelines; (7) promoting motivation, accountability and evaluation of skills; (8) access to care and (9) coordination of care. Eight of the nine priority areas were selected in 2003 after a thorough literature review, survey of international initiatives and consensus process among healthcare policy makers, healthcare professionals and consumers. The ninth was added in 2009. After defining healthcare priorities, QIs were identified within each area. Following literature review, 81 QIs were selected, and were then subjected to evaluation by healthcare professionals based on three dimensions: feasibility, importance, coherence with existing initiatives. A two-round Delphi method was used in order to select a first set of QIs to be developed and tested. In 2009, with the inclusion of a new priority area, this list was reviewed. QIs that were deemed to be ‘topped out’, or no longer priority measures, were discarded and new QIs were added. As a result, 42 QIs were selected for retooling and further development.16 Since 2006, 24 QIs developed by COMPAQH have been implemented nationwide (depending on the topic, among the 3000 HCOs, including 1300 acute-care organisations). Of the 24 QIs, 16 are based on PMRs, six on administrative data extracted from a national database, and two are based on an ad hoc survey (table 1).
Table 1

The 42 QIs developed for nationwide use in France

Type of hospitals concerned
Date of selectionPriority areaData sourceACRehab.PsyHome careIntroduced in*
(a) Indicators in nationwide use (n=24)
Consumption of antibiotics per 1000 patient-days20034AdminXXX2006
Composite index for evaluation of activities against nosocomial infections20034AdminXXX2006
Rate of surgical site infections (SURVISO)20034AdminX2006
Rate of methicillin-resistant Staphylococcus aureus per 1000 patient-days20034AdminXX2006
Annual volume of alcohol-based products (gels and solutions) per patient-day20034AdminXXX2006
Conformity of anaesthetic records20032PMRX2008
Delay in sending hospitalisation summary to general practitioner20039PMRX2008
Screening for nutritional disorders in adults20033PMRX2008
Medical record content20032PMRX2008
Traceability of pain assessment20031PMRX2008
Hospital care of myocardial infarction after the acute phase (8 QIs)†20036PMRX2008
Compliance of patient records in rehabilitation hospitals (4 QIs)20092PMRX2009
Traceability for risk assessment of pressure ulcers20096PMRXXX2009
Multidisciplinary meetings in oncology20032PMRX2010
Conformity of orders for imaging tests‡20032PMRXX2010
Compliance of patient records in homecare (5 QIs)20092PMRX2010
Compliance of patient records in psychiatry (3 QIs)20092PMRX2010
Prevention and management of postpartum haemorrhage (5 QIs)20096PMRX2012
Support for haemodialysis patients (X QIs)20096PMRX2012
Initial hospital treatment of stroke (6 QIs)20036PMRX2012
Satisfaction in hospitalised patients20035SurveyXIn progress
Waiting time for external consultation20038AdminXIn progress
Organisational support for breast cancer20036PMRXIn progress
Architectural, ergonomic and informational accessibility20038SurveyXXXIn progress
(b) Indicators in development (n=8)Priority areaData source
Organisational climate7Survey
Emergency timeout8Admin
Evaluation of patient complaints and claims5Admin
Detection of alcohol-dependent patients6PMR
Patient experience5Survey
Obesity surgery in adult6PMR
Composite score of professionals coordination on acute stroke management patients9Admin
Composite score of emergency department assessment8Admin
(c) Discarded indicators (n=10)Priority area
Absence of short-term professionals in contact with the patient7
Turnover of professionals in direct contact with the patient7
Cancellation of procedures involving anaesthesia in ambulatory care2
Violence in psychiatry4
Deadline for appointments in medico-psychological centres8
Management of treatment side-effects6
Electroconvulsive therapy6
Death in low-mortality diagnosis-related groups4
Hospitalised patients with a social management2
Prevention of falls in hospitalised patients4

*The year of national introduction. From the introduction, the QI is mandatorily reported each year by all hospitals concerned (except for ‘Conformity of orders for imaging tests’ QI which is not mandatory).

†Depending on the theme, one or more QIs were developed; we count 1 QI for 1 theme.

‡The only one that is not mandatory.

AC, acute care; Admin, administrative data-based; PMR, paper medical record; Psy, psychiatric care; QI, quality indicator; Rehab., rehabilitation care; Survey, ad hoc survey.

The 42 QIs developed for nationwide use in France *The year of national introduction. From the introduction, the QI is mandatorily reported each year by all hospitals concerned (except for ‘Conformity of orders for imaging tests’ QI which is not mandatory). †Depending on the theme, one or more QIs were developed; we count 1 QI for 1 theme. ‡The only one that is not mandatory. AC, acute care; Admin, administrative data-based; PMR, paper medical record; Psy, psychiatric care; QI, quality indicator; Rehab., rehabilitation care; Survey, ad hoc survey. Ten QIs were discarded because of low acceptance or poor metrological qualities.17 We shall use the results for the first six QIs to illustrate the methodology and the challenges encountered during development and implementation (table 2).
Table 2

Details of the first 6 QIs in nationwide use in France in acute-care hospitals

QINumber of records in random sample (n)*Calculation
Traceability of pain assessment80†Proportion of records containing at least one pain assessment result (number of records with at least one result/n)
Quality and content of the medical record80†Composite score (compliance with 10 items): presence of: surgical report, delivery report, anaesthetic record, transfusion record, outpatient prescription, outpatient record, admission documents, care and medical conclusions at admission, and drug prescriptions during stay; overall medical record organisation
Quality and content of the anaesthetic record60Composite score (compliance with 13 items). Presence of the following information:

Preanaesthesia: patient name, anaesthetist name, information on preanaesthesia visit, treatments, risk evaluation, type of anaesthesia, evaluation of access to upper airways

Peranaesthesia: anaesthetist name, technique of access to upper airways

Postintervention: anaesthetist name, discharge from recovery room, drug prescriptions

Perianaesthesia: adverse events

Time elapsed before sending discharge letters80†Proportion of records containing a letter sent to the patient's general practitioner within 8 days (number of records containing a letter/n)
Screening for nutritional disorders80†Proportion of records giving body weight (BW) at admission (number of records with BW/n)
Management of acute myocardial infarction at hospital discharge‡ (8 QIs)60Proportion of records1. With prescription for an antiplatelet drug (number with prescription+number justifying absence of prescription /n)2. With prescription for a beta-blocker (number with prescription+number justifying absence of prescription /n)3.1. With left ventricular ejection fraction (LVEF) measurement (number with LVEF/N)3.2. With LVEF <40% and prescription for an angiotensin-converting enzyme inhibitor (number of records with prescription/number of records with LVEF <40%)4.1. With prescription for a statin (number with prescription+number justifying absence of prescription/n)4.2. With prescription for a statin and order for lipid test (number of records with an order/number of records with a prescription)5. With advice on diet (number of records with advice/n)6. Of patients with a history of cigarette smoking who received advice on giving up (number of records with advice/number of records for patients with history of cigarette smoking)

*Previous year records for patients hospitalised for more than 1 day.

†Same sample used to measure 4 QIs.

‡Patients who died in hospital were excluded.

QI, quality indicator.

Details of the first 6 QIs in nationwide use in France in acute-care hospitals Preanaesthesia: patient name, anaesthetist name, information on preanaesthesia visit, treatments, risk evaluation, type of anaesthesia, evaluation of access to upper airways Peranaesthesia: anaesthetist name, technique of access to upper airways Postintervention: anaesthetist name, discharge from recovery room, drug prescriptions Perianaesthesia: adverse events *Previous year records for patients hospitalised for more than 1 day. †Same sample used to measure 4 QIs. Patients who died in hospital were excluded. QI, quality indicator.

Medical record sampling strategy

First, the QIs must be designed; a process that employs the collaboration of healthcare professionals and their representatives, in efforts to ensure face-to-face and content validation among stakeholders. After this stage, sufficient and accurate data need to be obtained for QI measurement. Our data collection method is based on manual data extraction from a random sample of 60–80 PMRs per hospital. The sample size needs to be small to limit workload and contain costs (feasibility). However, it also needs to be large enough to ensure reliability (ie, reproducible results for a fixed set of conditions irrespective of who makes the measurement) and discriminative power (ie, ability to detect overall poor quality and/or variations in quality among hospitals statistically).18 19 Discriminative power is crucial in order that public reporting and paying-for-performance mechanisms act as incentives for local quality improvement initiatives. Whenever possible, the same set of PMRs is used to measure several QIs.

QI feasibility, reliability and discriminative power

To ensure method validity, we assessed the metrological qualities of the QIs in a pilot test on 50 to more than 100 hospitals (depending on the QI). Three criteria were taken into account in the selection of hospitals: geographical area, volume of activity and status (teaching, public, private-non-profit and private-for-profit).

Feasibility

Poor data collection can diminish the metrological qualities of QIs. We used a 30-item standard grid to assess five dimensions of feasibility (acceptability, workload, understanding of QI implementation, professional involvement and organisational capacity including IT systems) in the pilot test.20–22 The grid was assessed in each pilot hospital by the health professional in charge of the data collection. In general, for each feasibility dimension, the total number of problems encountered was calculated. We considered a ‘no feasibility’ issue as under 5% of problems, ‘middle feasibility’ issue between 5% and 10% of problems, and a ‘feasibility’ issue up to 10% of problems. The feasibility problems encountered were then discussed in an expert panel before validating the content of each QI. In the pilot test, the highest incidence of feasibility problems was encountered with the acute myocardial infarction (AMI)-related QI (11% vs 4.9% for ‘quality and content of anaesthetic records’, and 3.2% or less for the other QIs). The dimension concerned was ‘professional involvement’. Originally, it was stipulated that a health professional of the specialty (eg, a cardiologist) must be involved in the data collection process, but during the pilot test, it was difficult to achieve this aim for all hospitals. The average time spent on data collection was 8.5 days per hospital for the five QIs related to medical or anaesthetic record content, and 5 days for AMI QIs (including sample of the medical records, retrieval from archives, abstraction of the sample, data entry in the computer and verification of data quality). As a result, a national generalisation committee was created, which meets each year, and whose discussions include difficulties, including feasibility, encountered by hospitals during generalisation. Currently, no intrinsic limitation of feasibility has been reported for this or any other QI during nationwide generalisation.23

Reliability

Interobserver reliability is essential for standardised manual data collection. The reliability of our method was tested by double-data capture of 20 PMRs by two independent observers in 10 hospitals. Observer agreement, as given by the Kappa coefficient, was in the range 0.80–0.96 for each QI.

Discriminative power

A QI should be able to detect room for improvement in clinical care. We routinely perform two tests: (1) the QI results obtained are first tested using Student's t test or Wilcoxon signed-rank test against an optimal threshold (100%). If the difference is significant (at 0.05%), the mean QI is below the desired threshold and (2) the dispersion of QI values among hospitals is tested using the Gini coefficient. This coefficient is a measure of statistical dispersion that is commonly used in economics to describe inequalities across groups.24 25 It is a ratio ranging from zero (maximum dispersion) to 1 (no dispersion). Table 3 gives the mean overall scores for each QI in our pilot study. Scores for all QIs varied widely across hospitals (Gini coefficient<0.5) except for ‘order of a lipid test’ (AMI 4.2 in table 3). This QI had a mean score (6.9%) far below the theoretical threshold of 100% (p<0.001). For each QI, performance (mean overall score) was significantly lower than the optimal threshold of 100% (p<0.001). Taken together, these results provide evidence for good discriminative power.
Table 3

Results for the 6 QIs during pilot testing

Score (%)
QIHospitals (n)MinMaxMean (SD)Gini coefficient
1. Quality and content of the medical record11236.989.964.2 (10.8)0.09
2. Screening for nutritional disorders128096.866.4 (22.6)0.2
3. Time elapsed before sending discharge letter13399152.7 (18.2)0.2
4. Letters Quality and content of the anaesthetic record8633.292.865.6 (12.4)0.1
5. Traceability of pain assessment133098.739.8 (27.8)0.4
6.1. AMI 1552010090.9 (11.9)0.05
6.2. AMI 25547.110084.6 (12.3)0.08
6.3. AMI 3.1554510087.7 (11.7)0.07
6.4. AMI 4.15572.710089.2 (6.9)0.04
6.5. AMI 4.255010014.5 (23.8)0.7
6.6. AMI 555093.333.7 (27)0.4

AMI, acute myocardial infarction; QI, quality indicator.

Results for the 6 QIs during pilot testing AMI, acute myocardial infarction; QI, quality indicator.

Hospital ranking

Our QI scores are computed together with their uncertainty, that is, the 95% CI around the mean score. The CI is estimated using the ‘Central limit theorem’ which assumes normal distribution of scores for item numbers above 30. The CI thus depends on the number of PMRs. Because of this, hospitals were ranked according to the Hospital Report Research Collaborative method into three categories (top, neutral, bottom) on the basis of the overall mean for all hospitals, and the 95% CI calculated for a normal distribution.26 Hospitals with fewer than 30 PMRs were excluded. Hospital distribution into the three categories was similar during the pilot test and after nationwide QI generalisation, except for 2 QIs, namely, ‘time elapsed since sending discharge letters’ and ‘screening for nutritional disorders’ (p<0.001) The difference was in favour of the pilot test (figure 1).
Figure 1

Variability in score for ‘quality and content of the medical record’, ‘traceability of pain assessment’ and ‘time elapsed before sending discharge letters’ during pilot testing. The horizontal line gives the mean score for each hospital (with 95% CI). The vertical line represents the overall mean score for all hospitals. The number and percentage of hospitals in each ranking category are given.

Variability in score for ‘quality and content of the medical record’, ‘traceability of pain assessment’ and ‘time elapsed before sending discharge letters’ during pilot testing. The horizontal line gives the mean score for each hospital (with 95% CI). The vertical line represents the overall mean score for all hospitals. The number and percentage of hospitals in each ranking category are given.

Discussion

We propose a method for developing QIs for nationwide hospital comparisons based on manual data extraction from PMRs. Our key concern is reducing the workload as far as possible without detracting from the validity of the statistical comparisons among hospitals. We achieved this by selecting the smallest PMR sample (60–80) that can discriminate among hospitals, by using the same set of PMRs to measure several QIs whenever possible, and by discarding QIs that had shown poor metrological qualities in a low-cost pilot test of feasibility, reliability and discriminative power.

Strengths of the method

Our method was adopted in France in 2006, and has led to the nationwide implementation of an increasing number of QIs. Public reporting of 1200 acute-care hospitals based on these QIs has been available since 2008 on the website of the French Ministry of Health.15 The method has three strengths: (1) the close collaboration of health professionals and their representatives in QI development, from the QI design and definition stage to feasibility testing, auditing and validation of any changes made, facilitates QI acceptability, appropriation of audit results and the introduction of quality improvement initiatives; (2) feasibility, reliability and discriminative power were assessed in a low-cost pilot test, which means that each year a new set of QIs can be prepared for generalisation and (3) QIs yielding poor results could be discarded before generalisation.

Limitations

Our study has limitations. The preset PMR sample size, regardless of the hospital's volume of activity, might introduce a bias. The bias and likelihood of an erroneous classification were limited by combining two approaches: (1) by introducing uncertainty (Van Dishoeck et al recently showed how, depending upon the method used, account can be taken of uncertainty27) and (2) by comparing three categories of hospitals rather than individual hospitals. A second limitation concerns the quality of the data in the PMR sample. Some gaming behaviour and some observer-dependent variability may persist in the data collection process. An ex-post control could be carried out in a sample of hospitals to address this limitation. A third limitation is the fact that this methodology can be time consuming, which limits the number of QIs nationally introduced each year. In order to mitigate this effect, we developed some specific strategies, such as having common collected data for multiple QIs, and a small sample of PMRs required per QI.

Its role in the context of development of EMR system

The above method is a powerful means of implementing an authoritative and valid national QI system, in both developed and developing countries, before EMRs become the norm. As attempts to adopt EMRs have faced several barriers, EMR introduction and adoption at a national level may well take longer than expected.12 In addition, the quality of the data that can be systematically extracted automatically is limited by both technical (IT system compatibility in hospitals) and ethical (confidentiality of information) obstacles to even partial EMR systems.28 Without hampering the development of EMRs and the meaningful use of health information technology, our PMR-based method represents a pragmatic alternative. Importantly, this method may be used in countries with complete paper records, as well as countries with mixed paper and electronic records. We recognise that even in nations with relatively high EMR adoption, often hospitals use both electronic and paper records. This methodology can be applied to the manual extraction of data from EMRs as needed, and act as a bridge to measuring quality in hospitals on a national level pending complete adoption of electronic systems. It is particularly well suited to practice guidelines, organisational issues related to the standardisation of work rules and coordination, and could also be used to assess the quality of patient care pathways.29
  22 in total

1.  From a process of care to a measure: the development and testing of a quality indicator.

Authors:  H R Rubin; P Pronovost; G B Diette
Journal:  Int J Qual Health Care       Date:  2001-12       Impact factor: 2.038

2.  Quality of health care for women: a demonstration of the quality assessment tools system.

Authors:  Elizabeth A McGlynn; Eve A Kerr; John Adams; Joan Keesey; Steven M Asch
Journal:  Med Care       Date:  2003-05       Impact factor: 2.983

3.  Research methods used in developing and applying quality indicators in primary care.

Authors:  S M Campbell; J Braspenning; A Hutchinson; M Marshall
Journal:  Qual Saf Health Care       Date:  2002-12

4.  Comparing clinical automated, medical record, and hybrid data sources for diabetes quality measures.

Authors:  Eve A Kerr; Dylan M Smith; Mary M Hogan; Sarah L Krein; Leonard Pogach; Timothy P Hofer; Rodney A Hayward
Journal:  Jt Comm J Qual Improv       Date:  2002-10

5.  Psychometric properties of the WHO Quality of Life questionnaire (WHOQOL-100) in patients with chronic diseases and their caregivers in China.

Authors:  Lingjiang Li; Derson Young; Shuiyuan Xiao; Xiao Zhou; Liang Zhou
Journal:  Bull World Health Organ       Date:  2004-07       Impact factor: 9.408

6.  Continuity of medical care: conceptualization and measurement.

Authors:  S M Shortell
Journal:  Med Care       Date:  1976-05       Impact factor: 2.983

7.  How many patients are necessary to assess test performance?

Authors:  C F Arkin; M S Wachtel
Journal:  JAMA       Date:  1990-01-12       Impact factor: 56.272

8.  Can performance-based financing be used to reform health systems in developing countries?

Authors:  Megan Ireland; Elisabeth Paul; Bruno Dujardin
Journal:  Bull World Health Organ       Date:  2011-09-01       Impact factor: 9.408

9.  Clinical governance in primary care groups: the feasibility of deriving evidence-based performance indicators.

Authors:  A McColl; P Roderick; H Smith; E Wilkinson; M Moore; M Exworthy; J Gabbay
Journal:  Qual Health Care       Date:  2000-06

10.  Measuring the quality of diabetes care using administrative data: is there bias?

Authors:  Nancy L Keating; Mary Beth Landrum; Bruce E Landon; John Z Ayanian; Catherine Borbas; Edward Guadagnoli
Journal:  Health Serv Res       Date:  2003-12       Impact factor: 3.402

View more
  6 in total

1.  Keys to successful implementation of a French national quality indicator in health care organizations: a qualitative study.

Authors:  Mathias Waelli; Marie-Léandre Gomez; Claude Sicotte; Adrian Zicari; Jean-Yves Bonnefond; Philippe Lorino; Etienne Minvielle
Journal:  BMC Health Serv Res       Date:  2016-10-06       Impact factor: 2.655

2.  Comparing routine administrative data with registry data for assessing quality of hospital care in patients with myocardial infarction using deterministic record linkage.

Authors:  Birga Maier; Katrin Wagner; Steffen Behrens; Leonhard Bruch; Reinhard Busse; Dagmar Schmidt; Helmut Schühlen; Roland Thieme; Heinz Theres
Journal:  BMC Health Serv Res       Date:  2016-10-21       Impact factor: 2.655

3.  Using standardised patients to assess the quality of medical records: an application and evidence from rural China.

Authors:  Yuju Wu; Huan Zhou; Xiao Ma; Yaojiang Shi; Hao Xue; Chengchao Zhou; Hongmei Yi; Alexis Medina; Jason Li; Sean Sylvia
Journal:  BMJ Qual Saf       Date:  2019-11-27       Impact factor: 7.035

4.  Assessing the impact of a quality improvement program on the quality and timeliness of discharge documents: A before and after study.

Authors:  Pénélope Troude; Isabel Nieto; Annie Brion; Raphaël Goudinoux; Jean Laganier; Valérie Ducasse; Rémy Nizard; Fabien Martinez; Christophe Segouin
Journal:  Medicine (Baltimore)       Date:  2020-12-18       Impact factor: 1.817

5.  Main barriers to effective implementation of stroke care pathways in France: a qualitative study.

Authors:  Kristel Gache; Henri Leleu; Gérard Nitenberg; France Woimant; Marie Ferrua; Etienne Minvielle
Journal:  BMC Health Serv Res       Date:  2014-02-28       Impact factor: 2.655

6.  Comparative Quality Indicators for Hospital Choice: Do General Practitioners Care?

Authors:  Marie Ferrua; Claude Sicotte; Benoît Lalloué; Etienne Minvielle
Journal:  PLoS One       Date:  2016-02-03       Impact factor: 3.240

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.