Literature DB >> 32519968

Developing and Applying a Formative Evaluation Framework for Health Information Technology Implementations: Qualitative Investigation.

Kathrin Cresswell1, Robin Williams2, Aziz Sheikh1.   

Abstract

BACKGROUND: There is currently a lack of comprehensive, intuitive, and usable formative evaluation frameworks for health information technology (HIT) implementations. We therefore sought to develop and apply such a framework. This study describes the Technology, People, Organizations, and Macroenvironmental factors (TPOM) framework we developed.
OBJECTIVE: The aim was to develop and apply a formative evaluation framework for HIT implementations, highlighting interrelationships between identified dimensions and offering guidance for implementers.
METHODS: We drew on an initial prototype framework developed as part of a literature review exploring factors for the effective implementation of HIT. In addition, we used qualitative data from three national formative evaluations of different HIT interventions (electronic health record, electronic prescribing, and clinical decision support functionality). The combined data set comprised 19 case studies of primarily hospital settings, and included 703 semistructured interviews, 663 hours of observations, and 864 documents gathered from a range of care settings across National Health Service (NHS) England and NHS Scotland. Data analysis took place over a period of 10 years and was guided by a framework informed by the existing evidence base.
RESULTS: TPOM dimensions are intimately related and each include a number of subthemes that evaluators need to consider. Although technological functionalities are crucial in getting an initiative off the ground, system design needs to be cognizant of the accompanying social and organizational transformations required to ensure that technologies deliver the desired value for a variety of stakeholders. Wider structural changes, characterized by shifting policy landscapes and markets, influence technologies and the ways they are used by organizations and staff.
CONCLUSIONS: The TPOM framework supports formative evaluations of HIT implementation and digitally enabled transformation efforts. There is now a need for prospective application of the TPOM framework to determine its value. ©Kathrin Cresswell, Robin Williams, Aziz Sheikh. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 10.06.2020.

Entities:  

Keywords:  evaluation; health information technology; sociotechnical

Mesh:

Year:  2020        PMID: 32519968      PMCID: PMC7315366          DOI: 10.2196/15068

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   5.428


Introduction

Health systems worldwide are prioritizing the implementation of health information technology (HIT) in the quest to address some of health care’s greatest challenges, including aging populations living with long-term conditions, persistent variations in the quality of care, and rising health care costs [1,2]. Although there is general agreement that HIT has the potential to improve safety, quality, and efficiency [3], large-scale HIT implementations require significant upfront investment, benefits are likely to materialize slowly, and those who put in most of the effort are often not those who benefit directly [4]. Such social and organizational challenges, which vary across contexts and technological functionalities, are often hard to navigate and predict for those managing change [5,6]. Despite efforts to identify success factors to guide implementation efforts, there is no simple recipe for success [7]. Formative evaluations can help to navigate these challenges. They can assist decision makers in moving from reactive to proactive strategies and identifying appropriate metrics to establish baselines and measure progress. In addition, they can help decision makers learn lessons more rapidly within the time frame of a project life cycle [8-10]. Formative evaluations can identify emerging unintended consequences and thereby, for instance, help to avert potential adverse outcomes for patient safety (eg, those arising from shortcomings in design, implementation strategy, and work practices) [11,12]. Evaluators should ideally work closely with strategic decision makers to keep projects on track and identify potential risks and mitigation strategies as early as possible. However, despite these potential benefits, there is limited expertise in health services in conducting such formative evaluations. Existing organizational implementation evaluations, if conducted at all, often take place after an implementation has gone wrong, and use suboptimal methodologies. Evidence-based frameworks to guide organizations in conducting HIT implementation evaluations have the potential to be helpful. A number of health informatics scholars have recently recognized this gap and developed various frameworks, some of which are summarized in Table 1 [13-19]. Some factors, such as user engagement and leadership, are well established in the change management literature. Other factors, such as political and market dimensions, have more recently received increasing recognition in shaping HIT implementations. The proliferation of frameworks poses a challenge for implementers seeking to navigate the literature and this paper seeks to integrate these frameworks.
Table 1

Examples of existing health information technology evaluation frameworks.

FrameworkKey characteristicsReference
Nonadoption, abandonment, scale-up, spread, and sustainability (NASSS) frameworkThis framework includes the following domains: the condition or illness, the technology, the value proposition, the adopter system, the organization(s), the wider context, and the interaction and mutual adaptation between all these domains over time.Greenhalgh et al [13]
Framework for Evaluation of Informatics ToolsThis framework includes the following stages: specification and needs requirements, component development, integration of system into a clinical setting, and routine use of a system.Kaufman et al [14]
Health Information Technology Evaluation ToolkitThis framework includes the following dimensions: articulating goals of the project, understanding stakeholders, and benefits measurement.Cusack and Poon [15]
Health Information Systems: human, organization, and technology-fit factors (HOT-fit)This framework focuses on the fit between technological, human, and organizational dimensions.Yusof et al [16]
Health Information Technology Reference-based Evaluation Framework (HITREF)This framework includes 6 dimensions: structural quality, functional quality, effects on quality processes, effects on outcome quality of care, unintended consequences, and barriers and facilitators.Sockolow et al [18]
Examples of existing health information technology evaluation frameworks. Based on over 10 years of experience, we set out to update current thinking about formative evaluation frameworks. Drawing on the existing literature, our aim was to develop and apply a formative evaluation framework for HIT implementations that would offer guidance for implementers. In this study, we will present our experiential conclusions and highlight interrelationships between identified dimensions.

Methods

Description of the Data Set

We have led a series of qualitative, theoretically informed case studies of different HIT implementations in the context of national formative evaluations. These included electronic health records (EHRs), clinical decision support (CDS) systems, and a combination of CDS and computerized physician order entry (CPOE) systems [4,20,21]. Our ongoing involvement as principal investigators and researchers in these various formative evaluations provided a platform for understanding HIT implementation challenges. Table 2 shows our data set, consisting of qualitative data collected between 2009 and 2018 by our research teams that included 11 social scientists. AS was the principal investigator on two of these projects [4,20], KC was the principal investigator on one [21], and RW was a senior adviser on all three [4,20,21]. We have published several papers from these evaluations, including both primary research and sets of evaluation recommendations based on the literature [4,20,21].
Table 2

Data set informing the development of the evaluation framework.

ProjectData setTimeline
National evaluation of the implementation of electronic health records in secondary care in England12 longitudinal qualitative case studies: 431 interviews, 590 hours of observations, 234 sets of field notes, and 809 documentsFebruary 2009 to January 2011
National evaluation of the implementation of clinical decision support/computerized physician order entry systems in English hospitals6 longitudinal qualitative case studies: 242 interviews, 32.5 hours of observations, and 55 documentsDecember 2011 to March 2016
National evaluation of a pilot decision support platform in Scottish primary care30 interviews and 8 nonparticipant ethnographic observationsMay 2018 to October 2018
KC collected some primary data (approximately 100 interviews and 60 hours of observations on [20], 40 interviews on [4], and 14 interviews on [21]). The majority of case study sites (18 of 19) included hospital settings implementing EHR and CDS/CPOE functionality in the English National Health Service (NHS). Data set informing the development of the evaluation framework.

Sampling Overview

We defined a case as an organization implementing relevant functionality within the boundaries of an organizational setting. We sampled hospitals for maximum variation in relation to geographical location, size, implementation strategy, technological systems, and governance structures [22]. We sampled individual participants through a combination of convenience and snowball approaches with key local gatekeepers facilitating initial contacts [23]. Participants in case study settings included representatives with varying degrees of seniority from a range of clinical professions (medical, nursing, pharmacy, and allied health care professionals), managerial, and IT support staff. We also collected data from relevant policymakers and system vendors, in order to gain insights into the wider market and policy dynamics in which local implementations took place.

Data Collection Overview

We collected data between February 2009 and October 2018. The majority of data consisted of digitally audio-recorded semistructured qualitative interviews (mainly face-to-face, some by telephone). These interviews explored expectations and experiences of implementing, using, and developing the new technology (depending on the background of the interviewee). Although interview guides varied with the specific focus functionality examined, key issues explored included the following: current systems, strategies, and organizational setup; views on potential system benefits and barriers to achieving these; and future directions and visions. In many cases (18 of 19 case studies), we sought to interview participants longitudinally (ie, before the implementation of the system, during the implementation, and once they had time to get used to the HIT system). Observations were nonparticipant in nature, opportunistic, and involved attending relevant strategic meetings (where the researcher took notes) or following a particular activity (eg, doctors using a specific system). Observations explored technological deployments in real-world contexts. Documents consisted of minutes of strategic meetings, summaries of lessons learned, and business cases. These provided insights into planned activities and local narratives surrounding implementation.

Development of the Framework

We began by conducting a review of existing frameworks and undertook a systematic literature review to explore which factors are important for the effective implementation of HIT [24]. The resulting prototype coding framework was iteratively refined over time and throughout projects. It included a number of dimensions and factors that formed the basis for coding qualitative data collected throughout case studies. In this process, we also allowed additional categories to emerge inductively [25]. Case studies were initially coded separately and then integrated iteratively across functionality (EHRs, CDS/CPOE in hospitals or what is known as ePrescribing in the United Kingdom, and CDS). This resulted in development of the prototype coding framework into a more comprehensive evaluation framework, which was synthesized to reflect the most pertinent categories and updated in light of the current literature. Here, our focus was on breadth rather than depth, aiming to produce a comprehensive overview of various stakeholder perspectives. We gave particular attention to stakeholders who were underrepresented (eg, vendors and administrative staff). Our analysis and development of the novel framework was informed theoretically by the sociotechnical approach, structuration theory, the social shaping of technology, and the theory of the diffusion of innovations [26-29]. The final framework was agreed upon through iterative discussion.

Results

We observed some commonalities across diverse settings and technological functions. The evaluation framework that has emerged from this work (Table 3, Figure 1) tackles important characteristics of the implementation landscape, where a range of technological, people (social/human), organizational, and wider macroenvironmental factors play an important role. Table 3 illustrates the key considerations in each of the dimensions. Figure 1 shows the interrelationship between the dimensions (technology, people, organizational, and macroenvironmental) and the various subcategories within each of these that need to be considered when implementing HIT.
Table 3

The Technology, People, Organizations, and Macroenvironmental factors (TPOM) framework, with example descriptions of dimensions.

Factor and dimensionDescription
Technological factors
UsabilityWhat is the ease of use and learnability of the technology?
PerformanceDoes the technology function as intended by developers?
Adaptability and flexibilityCan system design be changed to suit emerging needs?
DependabilityIs the system reliable and stable?
Data availability, integrity, and confidentialityIs data in the system available, accessible, and usable for those who need it?
Data accuracyIs the data in the system accurate?
SustainabilityIs use of the technology sustainable?
SecurityIs the system secure?
Social/human factors
User satisfactionWho are the users? Are users satisfied with the technology?
Complete/correct useAre features and functionality implemented and used as intended?
Attitudes and expectationsWhat benefits do users expect from using the technology and how can these be measured?
EngagementAre users actively engaged in implementation, adoption, and optimization?
ExperiencesDo users have negative experience with previous technologies?
Workload/benefitsAre the benefits and efforts relatively equal for all stakeholders?
Work processesDoes the system change relationships with patients, patterns of communication, and professional responsibilities (eg, increase of administrative tasks)?
User input in designIs there effective communication between designers, information technology staff, and end users, as well as between management and end users?
Organizational context
Leadership and managementAre management structures to support the implementation adequate?
CommunicationAre aims, timelines, and strategy communicated?
TimelinesAre implementation timelines adequate?
VisionWhat benefits do organizations expect from implementing the technology and how can these be measured? Is a coherent and realistic vision driving developments?
Training and supportIs the training adequate and realistic?
ChampionsAre champions and boundary spanners utilized?
ResourcesIs implementation adequately resourced? (includes technology, change management, and maintenance)
Monitoring and optimizationIs system performance and use monitored and optimized over time? Are lessons learned captured and incorporated in future efforts?
Wider macroenvironment
MediaHow is the technology viewed by the media and by the public? How does the organization view/manage media relations?
Professional groupsHow is the technology viewed by professional groups?
Political contextWhat benefits do policymakers expect from the technology and how can these be measured? What is the national approach to achieving interoperability and does the system align with this? Is there a coherent vision, consistent approach, and a clear direction of travel, allowing a degree of local input?
Economic considerations and incentivesAre there clear incentives for organizations and users to implement? (eg, improvements in quality of care) Is sufficient funding in place to support the initiative?
Legal and regulatory aspectsHave legal and regulatory frameworks been established?
VendorsIs vendor management effectively organized?
Measuring impactAre various stakeholders working together to define, validate, test, and refine outcome measures and measurement strategies? Are outcome measures important, clinically acceptable, transparent, feasible, and usable?
Figure 1

Diagram illustrating the evaluation framework.

The Technology, People, Organizations, and Macroenvironmental factors (TPOM) framework, with example descriptions of dimensions. Diagram illustrating the evaluation framework. None of these dimensions outranks the others. The relationship between dimensions influences how implementation, adoption, optimization, and maintenance processes unfold over time. Below, we illustrate some examples of the interrelationship between the different overarching TPOM dimensions (technology, people, organizational, and macroenvironment; Figure 1). The dimensions identified are not intended to provide detailed support for the different components of the framework, but rather to illustrate the relationships and interdependencies of the framework components.

Technological Factors and their Social Context of Use

We observed that, irrespective of the technology, systems needed to be usable, stable, and reliable (dependability), hold data securely (security), and only allow those with appropriate access rights to view confidential data (confidentiality). If systems and data within them were not dependable, users and implementers tended to lose confidence in their system choice. …we are now questioning whether [name] is the right solution for our high secure service… However, it was also apparent that the design of technologies never occurred in isolation of the social and organizational context of use. This was exemplified through the many different ways in which technologies transformed how users worked in often unanticipated ways (eg, by making data entry more cumbersome; see the Social/human dimension, which includes “User satisfaction,” “Workload/benefits,” and “Work processes,” in Table 3), but also by how different technologies were optimized to suit organizational and user requirements over time (“Adaptability” in Table 3). We had customized the system over a significant period of time to make it usable... System usability was the most important prerequisite for successful adoption. If users had to navigate a large number of interfaces and had difficulty finding relevant data, and if there was a general lack of intuitiveness, this slowed down their work (“Workload” and “Work processes” in Table 3). All our doctors and nurses are having to work harder now, because we are having to see the same number of patients with less time, because you are spending more time on a computer now. As a result, to not disrupt everyday delivery of care, staff had to employ workarounds (“Complete/correct use” in Table 3). The staff just create workarounds of the system and some of them are ingenious and have gone on to change the system in a good way, but some aren’t as helpful. Although changes to system design could address these issues, modifications were in many instances cumbersome and lengthy (“Adaptability and flexibility” in Table 3). Vendors, in turn, struggled to find a balance between tailoring of applications to local needs (achieved through local configuration) and developing commercially viable generic system versions (“Vendors” in Table 3). So when [vendor] deliver the product it will have a form designer so you will be able to go in the back end […]so the focus is to, for most of it make sure that it’s done through configuration not through software changes.

The Relationship Between Organizational Context and Macroenvironments

Organizations tended to employ two distinct strategies to approach system implementation. These involved both “top-down” and “bottom-up” management approaches, as well as technology strategies that involved transformative change combined with automating existing processes. “Top-down” and “bottom-up” strategies occurred at two different levels. These included the following: (1) technology design/procurement (in the English National Programme for IT (NPfIT) characterized by centralized procurement; “Political context” in the Wider macroenvironment dimension); and (2) technology implementation strategy within the organization (“Leadership and management” in the Organizational dimension). Balancing these tensions was instrumental for the perceived success of initiatives; this is not recognized in many existing frameworks. For example, although there was a perceived need for clinicians to be heavily involved in strategy, implementation, and deployment (“Communication” and “Vision”), stakeholders also recognized that some decisions at an organizational level had to satisfy the needs of diverse stakeholders, including those outside the immediate hospital environment (“Adaptability and flexibility” in Table 3). Everybody wants their own changes so some of the changes are the [management] have decided to do it that way but that doesn’t mean Consultant A thinks that’s the right way, so there is often not acceptance.

The Relationship Between Macroenvironments and Technological Factors

It was very clear throughout our work that wider structures had a significant impact on organizational processes, ways of working, and technologies (see the Macroenvironmental dimension in Table 3). These included tensions between the long time frames needed to achieve transformation (5 to 10 years) and the episodic funding schemes (2- to 3-year programs; “Timelines” in Table 3). We also observed changing policy landscapes that involved a high turnover in senior staff and accompanying changes in visions of digital care and available funding (see “Political context” in Table 3). For example, when we began our work, England’s national strategy of implementing centrally procured systems had just started [30]. However, during the CDS/CPOE evaluation, there was an increasing focus on local involvement in decision making, driven by the demise of the NPfIT and increased economic pressures resulting from the global recession. When we completed data collection, there was again a growing recognition that national guidance was crucial for promoting implementation progress and interoperability [31]. I’ve got a concern that if one of those two parties come into power and it seems highly likely that they will, that the National Programme might be closed and [system] might be shut down and what then happens…do they close the whole of the National Programme in which case, do we go back to where we were eight years ago? In addition, we observed shifts in market and vendor structures and the technologies themselves. These are not sufficiently accounted for in existing frameworks (see “Vendors” in the Wider macroenvironment dimension in Table 3). During the NPfIT, we observed a limited number of large vendors pushing other players out of the English/UK market, although the dynamics changed after the demise of NPfIT and the termination of associated contracts. This gave way to a more vibrant vendor landscape (although the market is still not very open to new vendors), which has important implications for organizations and users as they can only procure technologies from those that are currently available. With the breakdown of [NPfIT] we are now seeing a lot more [hospitals] looking to take advantage of electronic prescribing and we’re seeing an increased level of interest at this time to see if they can do this because effectively they’ve put the infrastructure in… We also saw how technologies were refined “in use” and through close working relationships between vendors and users over long periods (“Adaptability” in the Technology dimension in Table 3), which helped vendors/system designers overcome their limited knowledge of the social context of use. We have repeatedly seen the formation of vendor/user groups and observed how these helped to actively shape designs and markets (“User input in design”).

Discussion

Summary of Findings

We have developed an evaluation framework for implementers of HIT initiatives to guide implementation and optimization of functionality (Table 3). Although this draws on formative work, it can also guide summative evaluations. The TPOM framework includes key issues to consider in relation to technological, social/human, organizational, and wider macroenvironmental dimensions. Our ongoing work has shown that these dimensions are intimately related. Technologies never exist in isolation; it is therefore critical to appreciate that technological change will be accompanied by transformations in social groups, organizations, and the wider landscapes in which these are situated, including health policy, economic climates, and the development of markets.

Strengths and Limitations

This work has drawn on a substantive composite qualitative data set collected over a long time frame. Therefore, it helped us to assess which dimensions were relatively stable over time, and only these were included in the TPOM framework. Insights presented here are views derived following careful critical reflections on a series of evaluations. Although we acknowledge that many dimensions could be included, we deliberately attempted to keep the themes and subthemes manageable, thereby addressing a key issue in health technology evaluation: the usability of evaluation tools. We do not claim to capture all factors that play a role in implementation, adoption, and optimization of HIT, nor do we claim that our TPOM framework will provide a recipe for success. However, its pragmatic use in implementation and evaluation activity is likely to improve processes by prompting implementers to consider the most important dimensions influencing outcomes, thereby reducing unintended consequences and maximizing value. The framework now needs to be applied prospectively to confirm its utility across settings and regions. As such, we hope that it will provide a solid foundation for other countries to develop their own evaluation frameworks. A key challenge faced by most existing evaluation frameworks, including ours, is that they neglect to account for the dynamic relationship between social and technological dimensions of change. As this relationship is a process, it does not lend itself well to presentation in 2D pragmatic evaluation tools. We have illustrated this dynamism in the Results section, drawing on concrete examples.

Integration of Findings with the Current Literature

Many empirical studies of HIT implementation are primarily concerned with evaluating impact and therefore emphasize quantitative measurements guided by benefits realization frameworks [32,33]. Recent evaluation frameworks have expanded this limited focus to include a more in-depth appreciation of the interplay of social and technological factors shaping implementations. However, although acknowledging the complexity of the process, these nuanced frameworks tend to concentrate on one particular local and situated aspect of technology implementation, thereby neglecting the role of wider structuring conditions in shaping developments [16,34-37]. Others have considered wider structuring conditions, but the tools developed lack intuitiveness, usability, and practical applicability. For example, some frameworks that are designed to shed light on sociotechnical processes can become abstract and difficult to apply by those without academic backgrounds (which arguably includes the vast majority of those implementing change in health system settings) [38]. Others have become so complex that they may include a myriad of relevant dimensions, but this attempt to capture everything may result in a level of complexity that undermines the usefulness of the framework as a tool [13,26]. The challenge is to avoid Lewis Carroll’s cartographer’s dilemma of needing to make a comprehensive map on the same scale as the mapped landscape, which then no longer helps users navigate [39]. Our framework has sought to address this dilemma. There are several commonalities with existing frameworks, including a recognition of key technology, human and organizational dimensions, and their interrelationship [13-16,18]. However, TPOM is not condition- or illness-specific [13]. It is not concerned with the likelihood of the technology being adopted and its spread [13], but it is designed to help implementers of technology consider how implementation is progressing, the potential emerging risks, and what aspects therefore need attention to facilitate adoption. Implementers can apply TPOM to any HIT project at any stage of implementation. It builds on other evaluation frameworks that take into account the microcontext of use [14-16,18], to include consideration of wider macroenvironment dimensions that influence implementation and adoption. It is not concerned with management tools, but with alignment of perspectives [15]. As such, it does not provide a “recipe for success”; rather, it is a tool designed to help implementers navigate a complex landscape with many conflicting agendas and considerations. When problems or risks are identified with the help of the framework, these can be systematically targeted to facilitate implementation and adoption.

Policy Recommendations and Implications for Practice Emerging From This Work

Pragmatic formative evaluation frameworks can help to understand areas for potential improvement, benefits, and ways to streamline processes associated with technology implementation in health care settings. Evaluations need to move away from simple benefit realization approaches (that attempt to identify and measure benefits at the end of an implementation) toward formative evaluations that help the stakeholders involved adjust strategy along the way. Formative evaluations are part of a shift toward an evolutionary model where evaluation is a resource for faster and more effective learning. Our proposed framework is a guide for implementers of technological change initiatives, to assist in planning, or during implementations of HIT initiatives in health care settings. We invite those who use it to suggest changes in both content and usability, as this will help to maximize its use and application. In due course, we hope to be able to draw on a range of data collected through the lens of the framework in different settings, refine it, and develop new insights in relation to each of the dimensions. Large transformative policy programs aimed at facilitating technology implementation beyond hospitals are likely to require different evaluation frameworks, as their effects may be harder to trace and attribute.

Conclusions

We have drawn on a substantial body of data to develop the TPOM framework (Table 3), which stakeholders can use to monitor change processes and, if necessary, adjust the direction of HIT implementation projects. Going forward, a key challenge is likely to be the ongoing tension between attempts to capture the dynamics, processes, and interrelationships involved in technological change; the large number of these dimensions and their complexity; and the usability of evaluation tools by those delivering care, which is linked to their potential to have impact. We encourage prospective application of the TPOM framework to determine its value.
  26 in total

1.  Health technology assessment in its local contexts: studies of telehealthcare.

Authors:  Carl May; Maggie Mort; Tracy Williams; Frances Mair; Linda Gask
Journal:  Soc Sci Med       Date:  2003-08       Impact factor: 4.634

Review 2.  Organizational framework for health information technology.

Authors:  Helga E Rippen; Eric C Pan; Cynthia Russell; Colene M Byrne; Elaine K Swift
Journal:  Int J Med Inform       Date:  2012-02-28       Impact factor: 4.046

3.  Health Information Technology Evaluation Framework (HITREF) Comprehensiveness as Assessed in Electronic Point-of-Care Documentation Systems Evaluations.

Authors:  Paulina S Sockolow; Kathryn H Bowles; Michelle Rogers
Journal:  Stud Health Technol Inform       Date:  2015

4.  Launching HITECH.

Authors:  David Blumenthal
Journal:  N Engl J Med       Date:  2009-12-30       Impact factor: 91.245

Review 5.  Health information systems evaluation frameworks: A systematic review.

Authors:  Amirhossein Eslami Andargoli; Helana Scheepers; Diana Rajendran; Amrik Sohal
Journal:  Int J Med Inform       Date:  2016-10-15       Impact factor: 4.046

6.  Exploring the roots of unintended safety threats associated with the introduction of hospital ePrescribing systems and candidate avoidance and/or mitigation strategies: a qualitative study.

Authors:  Hajar Mozaffar; Kathrin M Cresswell; Robin Williams; David W Bates; Aziz Sheikh
Journal:  BMJ Qual Saf       Date:  2017-02-07       Impact factor: 7.035

7.  Workarounds to hospital electronic prescribing systems: a qualitative study in English hospitals.

Authors:  Kathrin M Cresswell; Hajar Mozaffar; Lisa Lee; Robin Williams; Aziz Sheikh
Journal:  BMJ Qual Saf       Date:  2016-04-29       Impact factor: 7.035

Review 8.  Review of health information technology usability study methodologies.

Authors:  Po-Yin Yen; Suzanne Bakken
Journal:  J Am Med Inform Assoc       Date:  2011-08-09       Impact factor: 4.497

9.  Implementation and adoption of nationwide electronic health records in secondary care in England: final qualitative results from prospective national evaluation in "early adopter" hospitals.

Authors:  Aziz Sheikh; Tony Cornford; Nicholas Barber; Anthony Avery; Amirhossein Takian; Valentina Lichtner; Dimitra Petrakaki; Sarah Crowe; Kate Marsden; Ann Robertson; Zoe Morrison; Ela Klecun; Robin Prescott; Casey Quinn; Yogini Jani; Maryam Ficociello; Katerina Voutsina; James Paton; Bernard Fernando; Ann Jacklin; Kathrin Cresswell
Journal:  BMJ       Date:  2011-10-17

10.  NHS Scotland's Decision Support Platform: a formative qualitative evaluation.

Authors:  Kathrin Cresswell; Margaret Callaghan; Hajar Mozaffar; Aziz Sheikh
Journal:  BMJ Health Care Inform       Date:  2019-05
View more
  8 in total

1.  Modeling Data Journeys to Inform the Digital Transformation of Kidney Transplant Services: Observational Study.

Authors:  Videha Sharma; Iliada Eleftheriou; Sabine N van der Veer; Andrew Brass; Titus Augustine; John Ainsworth
Journal:  J Med Internet Res       Date:  2022-04-21       Impact factor: 7.076

2.  Key Challenges and Opportunities for Cloud Technology in Health Care: Semistructured Interview Study.

Authors:  Kathrin Cresswell; Andrés Domínguez Hernández; Robin Williams; Aziz Sheikh
Journal:  JMIR Hum Factors       Date:  2022-01-06

3.  Benefits realization management in the context of a national digital transformation initiative in English provider organizations.

Authors:  Kathrin Cresswell; Aziz Sheikh; Bryony Dean Franklin; Susan Hinder; Hung The Nguyen; Marta Krasuska; Wendy Lane; Hajar Mozaffar; Kathy Mason; Sally Eason; Henry W W Potts; Robin Williams
Journal:  J Am Med Inform Assoc       Date:  2022-01-29       Impact factor: 4.497

4.  Driving digital health transformation in hospitals: a formative qualitative evaluation of the English Global Digital Exemplar programme.

Authors:  Marta Krasuska; Robin Williams; Aziz Sheikh; Bryony Franklin; Susan Hinder; Hung TheNguyen; Wendy Lane; Hajar Mozaffar; Kathy Mason; Sally Eason; Henry Potts; Kathrin Cresswell
Journal:  BMJ Health Care Inform       Date:  2021-12

Review 5.  Barriers and enablers to implementing and using clinical decision support systems for chronic diseases: a qualitative systematic review and meta-aggregation.

Authors:  Winnie Chen; Claire Maree O'Bryan; Gillian Gorham; Kirsten Howard; Bhavya Balasubramanya; Patrick Coffey; Asanga Abeyaratne; Alan Cass
Journal:  Implement Sci Commun       Date:  2022-07-28

6.  Successful integration of an automated patient-reported outcome measure within a hospital electronic patient record.

Authors:  Matthew T Neame; David Reilly; Ajmal Puthiyaveetil; Liza McCann; Kamran Mahmood; Beverley Almeida; Clare E Pain; Victoria Furfie; Andrew G Cleary
Journal:  Rheumatol Adv Pract       Date:  2022-08-17

7.  Evaluation Framework for Successful Artificial Intelligence-Enabled Clinical Decision Support Systems: Mixed Methods Study.

Authors:  Mengting Ji; Georgi Z Genchev; Hengye Huang; Ting Xu; Hui Lu; Guangjun Yu
Journal:  J Med Internet Res       Date:  2021-06-02       Impact factor: 5.428

8.  Promoting inter-organisational knowledge sharing: A qualitative evaluation of England's Global Digital Exemplar and Fast Follower Programme.

Authors:  Susan Hinder; Kathrin Cresswell; Aziz Sheikh; Bryony Dean Franklin; Marta Krasuska; Hung The Nguyen; Wendy Lane; Hajar Mozaffar; Kathy Mason; Sally Eason; Henry W W Potts; Robin Williams
Journal:  PLoS One       Date:  2021-08-02       Impact factor: 3.240

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.