Literature DB >> 33419909

Study protocol for a prospective, double-blinded, observational study investigating the diagnostic accuracy of an app-based diagnostic health care application in an emergency room setting: the eRadaR trial.

S Fatima Faqar-Uz-Zaman1, Natalie Filmann2, Dora Mahkovic3, Michael von Wagner4, Charlotte Detemble5, Ulf Kippke5, Ursula Marschall6, Luxia Anantharajah5, Philipp Baumartz5, Paula Sobotta5, Wolf O Bechstein5, Andreas A Schnitzbauer5.   

Abstract

INTRODUCTION: Occurrence of inaccurate or delayed diagnoses is a significant concern in patient care, particularly in emergency medicine, where decision making is often constrained by high throughput and inaccurate admission diagnoses. Artificial intelligence-based diagnostic decision support system have been developed to enhance clinical performance by suggesting differential diagnoses to a given case, based on an integrated medical knowledge base and machine learning techniques. The purpose of the study is to evaluate the diagnostic accuracy of Ada, an app-based diagnostic tool and the impact on patient outcome. METHODS AND ANALYSIS: The eRadaR trial is a prospective, double-blinded study with patients presenting to the emergency room (ER) with abdominal pain. At initial contact in the ER, a structured interview will be performed using the Ada-App and both, patients and attending physicians, will be blinded to the proposed diagnosis lists until trial completion. Throughout the study, clinical data relating to diagnostic findings and types of therapy will be obtained and the follow-up until day 90 will comprise occurrence of complications and overall survival of patients. The primary efficacy of the trial is defined by the percentage of correct diagnoses suggested by Ada compared with the final discharge diagnosis. Further, accuracy and timing of diagnosis will be compared with decision making of classical doctor-patient interaction. Secondary objectives are complications, length of hospital stay and overall survival. ETHICS AND DISSEMINATION: Ethical approval was received by the independent ethics committee (IEC) of the Goethe-University Frankfurt on 9 April 2020 including the patient information material and informed consent form. All protocol amendments must be reported to and adapted by the IEC. The results from this study will be submitted to peer-reviewed journals and reported at suitable national and international meetings. TRIAL REGISTRATION NUMBER: DRKS00019098. © Author(s) (or their employer(s)) 2021. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Keywords:  accident & emergency medicine; adult gastroenterology; adult surgery

Year:  2021        PMID: 33419909      PMCID: PMC7798704          DOI: 10.1136/bmjopen-2020-041396

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


This is the first prospective study to examine the diagnostic accuracy of an app-based diagnostic tool in an emergency room and the impact on clinical outcomes. The study will be conducted in a real-life setting to investigate the performance in a high stress environment and to provide rationale for routine clinical application. The double-blinded design will avoid bias regarding research findings. The primary limitation of an observational design is that only associations can be described, not causal relationships.

Introduction

Diagnostic errors, comprising inaccurate, delayed or missed diagnoses, are one of the major challenges in public healthcare.1 In the recent ‘Patient Safety Fact File’, WHO outlines 10 crucial facts about patient safety.2 Accordingly, adverse events are among the 10 leading causes of death and disability, contributing to approximately 10% of patients harmed during hospitalisation. Of note, 10%–20% of adverse events have been quoted to be particularly related to diagnostic failure, causing more harm to patients than medication or treatment errors.3–5 Further, false or delayed diagnoses are reported to be the most common reason for medical malpractice litigation.6 Graber et al estimated that diagnostic failures occurred in 5%–15% of cases, depending on the medical specialty with higher percentages assumed in primary care and emergency medicine.7 Various reasons have been identified to contribute to false diagnoses. Graber concluded that cognitive slips, primarily resulting from faulty information processing and verification, and misguided situational confidence occur most frequently.8 9 This is especially evident in ER settings, which often have to deal with high throughputs, fast decision making and incomplete clinical information in a disruptive environment. In particular, emergency room (ER) overcrowding has been identified as a serious threat to patient safety, resulting in poor clinical outcome and a significant increase in mortality.10 Previous studies have revealed that more than 40% of admission diagnoses at first presentation to the ER are not concordant with the final diagnosis of the patient.11–13 That means, that throughout the hospital stay, the patient experiences a change in diagnosis based on a variety of additional diagnostics and reevaluation of initial assumptions, finally leading to the correct diagnosis. In particular, approximately 30% of patients with abdominal pain, being one of the leading causes for visiting the ER, exhibit a discrepancy in diagnosis.14 15 In particular, misdiagnosis rate of acute appendicitis, the most frequent reason for acute abdominal pain, has largely remained unchanged over time and is still associated with a high ratio of negative appendectomies.16 Inaccurate diagnosing in ERs has been shown to be further associated with increased length of hospital stay, rate of consultations, healthcare cost, and risk for mortality and morbidity, contributing to a serious concern to patient safety.11 13 17 18 Thus, a high degree of diagnostic accuracy can lead to an improvement in quality of patient care. Correct admission diagnoses are crucial for a reliable triage and process management and critically influence the initial evaluation in that ER and subsequent clinical course of the patient.19 Digital technologies and artificial intelligence (AI)-based methods have recently emerged as impressively powerful tools to empower physicians in clinical decision making and improve healthcare quality. More specifically, diagnostic decision support systems (DDSS) have demonstrated to facilitate assessment of clinical data input by using an extensive medical knowledge base.20 21 One version of DDSS is Ada, an app-based AI-machine learning system that incorporates patients’ symptoms and other findings into its knowledge base and intelligent technology to deliver effective healthcare.22 23 Based on an algorithmic pathway and driven by chief complaints, the app-based system generates a set of differential diagnoses for a given clinical case. Several studies have reported that DDSS have the potential to increase diagnostic performance, obtaining an accuracy rate of 70%–96%.24 25 In particular, a retrospective study of rare diseases has demonstrated that Ada suggests accurate diagnoses earlier than clinical diagnoses in more than half of all cases.23 However, application of the Ada app has not been investigated in a real-life setting, particularly in ERs, which has to deal with a high stress environment and heavy time constraints. This app-based method may be a valuable companion in triaging patients and support clinicians in making decisions more accurate and sooner by simultaneously reducing risk for medical errors. Therefore, in the present study, we aim to evaluate the diagnostic ability of Ada in ER settings and examine the impact on timing of diagnosis.

Methods and analysis

The eRadaR-trial is designed as a prospective, double-blinded, observational study evaluating the diagnostic accuracy of the Ada-App in the ER of the Department of General, Visceral and Transplant Surgery of the Frankfurt University Hospital, Germany. The trial protocol is written in accordance with the current Standard Protocol Items: Recommendations for Interventional Trials (SPIRIT 2013). The SPIRIT checklist is given in online supplemental Additional file 1.

The Ada-App specifications and rationale to use the software

The Ada-App is a class I medicinal product certified in accordance with the DIN ISO 13485. Ada is a free-downloadable certified medicinal product and has been validated in different studies by the marketing authorisation holder and developer team. It has shown a higher accuracy (73%) in comparison to other apps (38%) when compared with the correctness of symptom checking. The App was superior to other apps when the hitlist of the five most probable diagnoses were compared (84% vs 51%).22 26–32 The evidence shows that the algorithm is superior to other solutions on the market, it has been validated by the company, and the data were the basis for the certification as a medicinal product class I (CE-mark in accordance with DIN ISO 13485), supporting our rationale to test the potentially most beneficial and promising software on the market.

Study population and eligibility criteria

All patients presenting to the ER with abdominal pain will form the study population and be screened for trial eligibility. Notably, patients presenting with abdominal pain as part of multiple chief complaints (eg, chest pain and abdominal pain) will also be included in the study. Moreover, patients, who will be immediately discharged from the ED on the same day and patients, who will be admitted to the hospital after presenting to the ER will be both included in the study and followed up in an intention-to-treat fashion. Inclusion criteria comprise: (1) adults aged ≥18 years, (2) patients presenting with abdominal pain to the ER and (3) patients willing to participate and able to provide written informed consent. The criteria of exclusion are: (1) intubated patients, (2) unstable patients or (3) patients with severe injuries requiring immediate medical treatment, (4) patients unwilling or incapable of providing informed consent. Eligible patients are asked for their participation in the trial and written informed consent will be obtained from themselves. All reasons for exclusion of patients will be recorded in the trial screening log and analysed accordingly.

Description of study visits and assessment schedule

Eligible patients will be interviewed by the study team with the Ada-App based on an algorithmic pathway of questions relating to the symptoms. The Ada-App will only obtain data about patient demographics, patient history and information about current complaints. Patient’s name and date of birth will be pseudonymised using an individual identification code, as described in the section ‘data management and data safety’. Throughout the study, the patient, the study team and the physician treating the patient will be blinded regarding the list of proposed diagnoses by the app. The patient will subsequently be diagnosed by classical doctor-patient interaction and decision making. The clinical course of the patient will be followed until day 90 after initial contact in the ER. Detailed information about outline of the study and assessment schedule are displayed in table 1 and figure 1.
Table 1

Schedule of study visits and assessments of the eRadaR study

BaselineHospital stay*Discharge90 days FU
Visits V1 (Day 0) V2 and 3 (Days 7, 14) V4 V5 (Day 90)
Informed consentX
Eligibility criteriaX
Demographic data:XX
 (A) CCIX
 (B) RAI-C score
Ada diagnosis listX
ICD-10 diagnosesXXX
SymptomsX
Diagnostics†XX
Therapy and OPS-codeXX
Rate of consultationsX
Complications (CCI)XXX
Length of hospital stayX
Overall survivalX

(A) Routine blood samples (C reactive protein, white cell count, haemoglobin, platelets, sodium, potassium, creatinine, albumin, bilirubin, International normalized ratio (INR) (B) instrumental diagnostics (ultrasound, chest/abdominal CT/MRI, ECG, endoscopy).

*Visit 2 or 3 is left out, if the patient is discharged before.

†Diagnostics include.

CCI, Comprehensive Complication Index; FU, follow-up; ICD, International Classification of Diseases; OPS, operations and procedures; RAI-C, Risk Analysis C score; V, visit.

Figure 1

Study flow chart of the eRadaR study. CCI, Comprehensive Complication Index.

Schedule of study visits and assessments of the eRadaR study (A) Routine blood samples (C reactive protein, white cell count, haemoglobin, platelets, sodium, potassium, creatinine, albumin, bilirubin, International normalized ratio (INR) (B) instrumental diagnostics (ultrasound, chest/abdominal CT/MRI, ECG, endoscopy). *Visit 2 or 3 is left out, if the patient is discharged before. †Diagnostics include. CCI, Comprehensive Complication Index; FU, follow-up; ICD, International Classification of Diseases; OPS, operations and procedures; RAI-C, Risk Analysis C score; V, visit. Study flow chart of the eRadaR study. CCI, Comprehensive Complication Index.

Patient presenting to the ED (visit 1)

After enrolment in the trial, a structured interview with the Ada-App will be conducted and baseline data will be assessed including demographic data according to the Carlson Comorbidity Index and the Risk Analysis Index-C score (RAI-C score), the patients’ symptoms and International Classification of Diseases 10th Revision (ICD-10) diagnoses list.33–35 Participants are then diagnosed and treated according to the standard of care by the attending physician of the ER. As this is a double-blinded study to patients and treating physicians, Ada-App diagnoses lists will be randomly allocated to a study-ID and then manually transferred into the electronic case report forms (eCRF). The trial personnel will be blinded until the end of the study to avoid bias regarding subsequent diagnoses and treatment of the patient, except of the interim analysis, which is mentioned in the section of statistical analysis.

Hospital stay (visit 2, day 7)

This visit is performed on day 7, after the patient is admitted to the hospital. Data about diagnostics and therapies are assessed comprising laboratory results (ie, C reactive protein, white cell count, platelets, haemoglobin, bilirubin, creatinine, sodium and potassium, albumin, INR), computer-assisted diagnostics (ie, ultrasound, chest/abdominal CT/MRI, ECG, endoscopy), type of therapy (conservative, interventional or surgery), OPS code of therapies and complications according to the Comprehensive Complication Index (CCI) together with the date of occurrence.36 If the patient has not been admitted to the hospital or is discharged before day 7, visit 2 is left out.

Hospital stay (visit 3, day 14)

Visit 3 is performed on day 14 after patient’s admission and assessment schedule is equivalent to visit 2. If the patient has not been admitted to the hospital or is discharged before day 14, visit 3 is left out.

Discharge (visit 4)

At discharge, data including the final ICD-10 diagnosis and the timing of diagnosis will be recorded to subsequently analyse the accuracy and the timing of the Ada-App compared with the classical doctor–patient encounter. Further data items include diagnostics (laboratory, instrumental), OPS codes and type of therapies, complications according to CCI, length of hospital stay, overall health cost, rate of consultation.

Follow-up (visit 5)

The follow-up will be performed as a structured telephone interview or in person on day 90 and will encompass following data items: demographic data according to the RAI-C score, complication assessment according to CCI and overall survival.

Interventions

As this is an observational, double-blinded, prospective study, no experimental or control interventions are conducted.

Endpoints

Primary endpoint

The primary endpoint of this study is to evaluate the diagnostic accuracy of the Ada-App by comparing the decision making of the classical doctor–patient interaction with the diagnoses proposed by the app-based algorithm.

Secondary endpoint

Secondary endpoints of this study consist of the following: timing of final discharge diagnosis and time to treatment during hospital stay, comparing accurate diagnoses with discharge diagnoses as descriptive assessments, the occurrence of complications according to the CCI, total length of stay in hospital from initial contact in the ER until discharge, patient morbidity and mortality at day 90, overall health cost analysis and consultation rate. Further endpoints are displayed in the description of assessment schedule (table 1).

Measurement methods

For data capture, following measurement methods will be used: Primary outcome measurement will be performed using the Ada-App which will deliver a set of differential diagnoses to a given clinical case.23 Based on an algorithmic questionnaire and machine learning technologies, the Ada chatbot assesses symptoms of the patient, similar to the anamnestic techniques and clinical reasoning of physicians. Patients’ data are integrated into an extensive knowledge base, which has been specifically designed by medical doctors by incorporating validated disease models and comprehensive medical literature. Then, differential diagnoses are generated and ranked in order considering two features: the probability, based on epidemiological data and the best match between the diagnosis and the given symptoms. Through AI-based methods and multiple feedback loops, the Ada knowledge base grows after each interaction and diagnostic ability improves continuously. The occurrence of complications as secondary outcomes will be evaluated and analysed according to the CCI.36 The CCI represents the standard assessment of postoperative morbidity and comprises all complications occurring during a patient’s course based on the Clavien-Dindo classification (CDC). Compared with the CDC, which ranks complications based on the severity of the therapeutic consequence and grades them in five levels, the CCI uses a formula to integrate all complications, ranging them from 0 (‘no complication’) to 100 (‘death’).37 This advanced approach enables comparison of patients harbouring more than one complication and takes more subtle differences into consideration. For assessment of comorbid diseases and frailty-associated risk in a surgical population, we will use the Charlson Comorbidity Index and the RAI-C score.

Risk–benefit assessment

This is an observational, non-interventional study and does not comprise any specific risk for the patient, as data obtained with the app are not used in the ER standard of care. Therefore, there is no special need for additional safety management. A delay in the diagnosis and treatment of patients presenting to the ER is not expected, as the app-based interview will not require more than 10 min and will exclusively be performed in the waiting zone of the ER by the study team. Baseline assessment (during visit 1) will directly be conducted after patient has been registered at the ER and given informed consent. Besides that, unstable patients requiring immediate medical care are excluded from the study beforehand.

Data management and data safety

The investigators will design and produce eCRF for protocol-required data collection. All information will be entered into these eCRFs by authorised and trained members of the study team and systematically checked for accuracy and completeness. Staff members with responsibilities for data collection or those, having access to the database will be enrolled in a delegation log. Patients’ data collected during the trial will be recorded in pseudonymised form by solely using individual identification codes. For data assessment using the Ada-App, a specified iPad will be provided, which will be registered at the Frankfurt University Hospital and will be exclusively used for the purpose of this trial. Clinical data will be documented pseudonymously by using a combination of a random number from 1 to 450 and the patient’s year of birth. Participants are then asked to answer the questionnaire of the Ada-App preferably by themselves or otherwise assisted by the study team. The diagnoses will be manually transferred into the eCRF of the related patient after trial completion and unblinding. All trial data obtained will be integrated into a statistical analysis software and analysed by the Institute of Biostatistics and Mathematical Modelling Frankfurt.

Ethics and dissemination

The eRadaR trial will be conducted in accordance with the Declaration of Helsinki and the international conference of harmonisation good clinical practice guidelines. After a patient has been identified to meet eligibility criteria, the patient will be informed about the aim, outline and individual risk of the study and informed consent will be given. After a sufficient period, the patient can then sign informed consent and will receive a signed copy. The results of this trial will be submitted for publication in a peer-reviewed journal in a summarised anonymised manner. The study is scientifically supported by the Barmer health insurance company. Barmer will act as a scientific advisor regarding the conduct of the study, will be involved in the process of interpreting the data and in the publication and public distribution process of the study after trial completion. However, there will be no raw data sharing or financial support from the institution.

Statistical analysis

Interim analysis

One formal unblinded interim analysis of the trial data is planned to be performed after enrolment of about 200 patients to evaluate the diagnostic accuracy of the Ada-App with 90 days follow-up information. Statistical analysis will be performed by the responsible study biometrician using a significance level of alpha=0.001 and a subsequent report will be written. These results will be discussed with the investigators and the study team in a staff meeting and the continuation of the trial will be considered.

Sample size calculation and study duration

The assumptions that were made, was that more than 30% of the admission diagnoses are not consistent with the final discharge diagnosis and hypothesised that the Ada-App will increase the diagnostic accuracy from 70% to a rate of 85%. Providing a power of 90% and a two-sided significance level of alpha 5%, a target sample size of N=405 patients has to be recruited to detect the targeted effect. With an estimated dropout rate of 10%, we plan to recruit N=450 patients in this trial. Furthermore, we expect the width of the confidence intervals for the diagnostic accuracy to be 0.1 at maximum (0.09 with an estimated diagnostic accuracy of 0.7, 0.07 with an estimated diagnostic accuracy of 0.85). This trial is anticipated to start in September 2020 and the duration of patient’s participation is 3 months including follow-up. To achieve the required sample size of patients, trial completion is expected to be in 12 months (August 2020).

Patient and public involvement

Patients were not involved in the development of the research question or study design. They will, however, be involved in visit 1 and will be interviewed by the study team using the Ada-App. Further, the follow-up (visit 5) will be performed as a telephone interview or in person with the patients for data assessment.

Discussion

Diagnostic errors have been identified as a serious threat to patient safety, leading to preventable adverse events, particularly in ERs with a disruptive environment. AI-based tools and algorithms have the potential to substantially reduce diagnostic failures, achieving high rates of diagnostic accuracy, which rivals the capability of clinicians. A previous study provides an overview of the main types of existing tools, which are classified into categories related to the targeted step of diagnostic processing.25 Over the past few decades, a number of computerised DDSS have been developed, exhibiting promising diagnostic efficacy. Bond et al evaluated four current DDSS using clinical cases from the New England Journal of Medicine, demonstrating that Isabel and Dxplain achieve the strongest performance.38 Compared with former programmes, second-generation DDSS are far more powerful, providing more accurate suggestions with increasing complexity, while concomitantly requiring less time for diagnosing.21 24 This is primarily essential in an era of ER crowding, where fast and accurate triaging is necessary to prioritise critically ill patients and to optimise resource allocation.8 Stewart et al recently summarised various fields of AI application becoming relevant in emergency medicine, including imaging, decision-making, and outcome prediction.39 In terms of triaging, a machine learning-based tool efficiently predicts critical patient outcome, equivalent to the classically used Emergency Severity Index.40 In a prospective, multicentre study, the DDSS Isabel achieved high accuracy in diagnosing patients presenting to the ER, suggesting the final discharge diagnosis in 95% of cases.41 Another clinical decision support system has been evaluated in patients presenting with acute abdominal pain aiming to identify high-risk patients for acute appendicitis.42 Based on automated methods and an integrated risk calculator, patient data was assessed from the electronic health record (EHR) and management strategies suggested according to the risk level. Incorporation into EHR represents one of the most recent advances in the development of DDSS using ‘natural language processing’ techniques, which matches entered clinical data with the underlying knowledge base.43 This might facilitate assessment of larger volumes of data, save more time, and might increase acceptance of DDSS in clinical workflow. However, in most of these trials using clinical support systems, impact on patient outcome, or on healthcare costs were not assessed. Although diagnoses suggested by DDSS mostly contained the correct diagnosis and achieved high level of users’ satisfaction, relevance and specificity of extensive lists were low.20 25 38 Long lists may lead to distraction or to unnecessary diagnostic with increased risk for iatrogenic injuries and costs. In general, despite the given potential efficacy of DDSS, widespread acceptance for implementation of DDSS into the routine clinical practice is evolving scarcely.44 Studies focusing on AI-based diagnostic tools are generally designed heterogeneously and are often of poor quality, making it difficult to recommend widespread evidence-based clinical application.21 25 While most of the current trials demonstrated high diagnostic accuracy in retrospective and simulated cases, only few studies evaluated their performance in real clinical settings, particularly in high stress environments like ERs. Thus, further validations in prospective studies are required to investigate the diagnostic efficiency and utility of DDSS and their impact on routine clinical decision-making and patient outcome.
  39 in total

1.  Claims, errors, and compensation payments in medical malpractice litigation.

Authors:  David M Studdert; Michelle M Mello; Atul A Gawande; Tejal K Gandhi; Allen Kachalia; Catherine Yoon; Ann Louise Puopolo; Troyen A Brennan
Journal:  N Engl J Med       Date:  2006-05-11       Impact factor: 91.245

2.  Validation of a diagnostic reminder system in emergency medicine: a multi-centre study.

Authors:  Padmanabhan Ramnarayan; Natalie Cronje; Ruth Brown; Rupert Negus; Bill Coode; Philip Moss; Taj Hassan; Wayne Hamer; Joseph Britto
Journal:  Emerg Med J       Date:  2007-09       Impact factor: 2.740

3.  The Clavien-Dindo classification of surgical complications: five-year experience.

Authors:  Pierre A Clavien; Jeffrey Barkun; Michelle L de Oliveira; Jean Nicolas Vauthey; Daniel Dindo; Richard D Schulick; Eduardo de Santibañes; Juan Pekolj; Ksenija Slankamenac; Claudio Bassi; Rolf Graf; René Vonlanthen; Robert Padbury; John L Cameron; Masatoshi Makuuchi
Journal:  Ann Surg       Date:  2009-08       Impact factor: 12.969

4.  Development and Initial Validation of the Risk Analysis Index for Measuring Frailty in Surgical Populations.

Authors:  Daniel E Hall; Shipra Arya; Kendra K Schmid; Casey Blaser; Mark A Carlson; Travis L Bailey; Georgia Purviance; Tammy Bockman; Thomas G Lynch; Jason Johanning
Journal:  JAMA Surg       Date:  2017-02-01       Impact factor: 14.766

5.  Disagreement between emergency department admission diagnosis and hospital discharge diagnosis: mortality and morbidity.

Authors:  James Eames; Arie Eisenman; Richard J Schuster
Journal:  Diagnosis (Berl)       Date:  2016-03-01

6.  Analysing potential harm in Australian general practice: an incident-monitoring study.

Authors:  A L Bhasale; G C Miller; S E Reid; H C Britt
Journal:  Med J Aust       Date:  1998-07-20       Impact factor: 7.738

Review 7.  Use of health information technology to reduce diagnostic errors.

Authors:  Robert El-Kareh; Omar Hasan; Gordon D Schiff
Journal:  BMJ Qual Saf       Date:  2013-07-13       Impact factor: 7.035

8.  Can a decision support system accelerate rare disease diagnosis? Evaluating the potential impact of Ada DX in a retrospective study.

Authors:  Simon Ronicke; Martin C Hirsch; Ewelina Türk; Katharina Larionov; Daphne Tientcheu; Annette D Wagner
Journal:  Orphanet J Rare Dis       Date:  2019-03-21       Impact factor: 4.123

9.  Emergency department crowding: A systematic review of causes, consequences and solutions.

Authors:  Claire Morley; Maria Unwin; Gregory M Peterson; Jim Stankovich; Leigh Kinsman
Journal:  PLoS One       Date:  2018-08-30       Impact factor: 3.240

10.  Syndromic Surveillance Insights from a Symptom Assessment App Before and During COVID-19 Measures in Germany and the United Kingdom: Results From Repeated Cross-Sectional Analyses.

Authors:  Alicia Mehl; Francois Bergey; Caoimhe Cawley; Andreas Gilsdorf
Journal:  JMIR Mhealth Uhealth       Date:  2020-10-09       Impact factor: 4.773

View more
  2 in total

1.  Design and Implementation of a Comprehensive AI Dashboard for Real-Time Prediction of Adverse Prognosis of ED Patients.

Authors:  Wei-Chun Tsai; Chung-Feng Liu; Hung-Jung Lin; Chien-Chin Hsu; Yu-Shan Ma; Chia-Jung Chen; Chien-Cheng Huang; Chia-Chun Chen
Journal:  Healthcare (Basel)       Date:  2022-08-09

Review 2.  Large Bowel Ischemia/Infarction: How to Recognize It and Make Differential Diagnosis? A Review.

Authors:  Francesca Iacobellis; Donatella Narese; Daniela Berritto; Antonio Brillantino; Marco Di Serafino; Susanna Guerrini; Roberta Grassi; Mariano Scaglione; Maria Antonietta Mazzei; Luigia Romano
Journal:  Diagnostics (Basel)       Date:  2021-05-30
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.