| Literature DB >> 15078604 |
Dena M Bravata1, Vandana Sundaram, Kathryn M McDonald, Wendy M Smith, Herbert Szeto, Mark D Schleinitz, Douglas K Owens.
Abstract
We evaluated the usefulness of detection systems and diagnostic decision support systems for bioterrorism response. We performed a systematic review by searching relevant databases (e.g., MEDLINE) and Web sites for reports of detection systems and diagnostic decision support systems that could be used during bioterrorism responses. We reviewed over 24,000 citations and identified 55 detection systems and 23 diagnostic decision support systems. Only 35 systems have been evaluated: 4 reported both sensitivity and specificity, 13 were compared to a reference standard, and 31 were evaluated for their timeliness. Most evaluations of detection systems and some evaluations of diagnostic systems for bioterrorism responses are critically deficient. Because false-positive and false-negative rates are unknown for most systems, decision making on the basis of these systems is seriously compromised. We describe a framework for the design of future evaluations of such systems.Entities:
Mesh:
Year: 2004 PMID: 15078604 PMCID: PMC3322751 DOI: 10.3201/eid1001.030243
Source DB: PubMed Journal: Emerg Infect Dis ISSN: 1080-6040 Impact factor: 6.883
Summary of evaluation data for detection systems and diagnostic decision support systems for a bioterrorism response
| Evaluation criteria | Detection systems evaluated % (yes/total) | Diagnostic decision support systems evaluated % (yes/total) |
|---|---|---|
| Is the timeliness of diagnostic information described? | 36 (20/55) | 48 (11/23) |
| Are diagnostic sensitivity and specificity described? | 1.8 (1/55) | 13 (3/23) |
| Is the reference standard against which the system was compared described? | 7 (4/55) | 39 (9/23) |
| Are the system’s security measures described? | 0 | 0 |
| Is the evaluation of the system over a range of clinical situations or patient populations described? | 0 | 0 |
| Is the portability of the system described? | 54 (15/28) | NAa |
| Is the system’s ability to run more than one sample at a time described? | 10 (4/41) | NA |
| Is the system’s ability to detect more than one bioterrorism agent described? | 32 (12/37) | NA |
| Is the system’s ability to detect either/both toxins and organisms described? | 5 (2/37) | NA |
| Is the inclusion of all bioterrorism agents and associated illnesses in the system’s knowledge base described? | NA | 26 (5/19) |
| Is the flexibility to update the probability of bioterrorism-related illness as the epidemic progresses described? | NA | 0 |
| Is the method of reasoning used by inference engine described? | NA | 26 (5/19) |
| Is the use of standard vocabulary described? | NA | 0 |
aNA; not applicable.
Evaluation data for detection systems for bioterrorism agentsa
| System name | Purpose | Evaluation datab |
|---|---|---|
| Anthrax Sensor | A portable detection system for “highly sensitive detection of biological agents within seconds” | Reported to be capable of detecting endotoxins at a level that is “20 times lower than previously achieved by similar devices” |
| BioCapture | A portable collection system for use by first responders. | Was compared to an All Glass Impinger (AGI) that collects samples into liquid and a Slit Sampler that impacts bacteria directly onto growth media and found to have a collection efficiency of 50%-80% relative to the AGI and 60%-125% relative to the Slit Sampler |
| Digital Smell/Electronic Nose | To detect and classify microorganisms according to the volatile gases given off during metabolism. | An array of 15 sensors was able to correctly classify 68 of 90 colonies containing 0 or 1 of 5 test organisms and an uninoculated control; however, it registered 22 of 90 as false-positives |
| Fluorescence-based array immuno-sensor | To provide simultaneous, antibody-based detection of bioactive analytes in clinical fluids. | Bioterrorism agents intended to be detected include |
| LightCycler; Ruggedized Advanced Pathogen Identification Device (RAPID) | LightCycler uses a PCR cycler for “real-time” quantification of DNA samples. RAPID is a rugged, portable system that uses LightCycler technology for field detection of bioterrorism agents. | RAPID is reported by the manufacturer to be |
| MiniFlo | For rapid, portable detection of multiple biological agents using flow cytometry. | Detected 87% of unknown biological agent simulants, including agents similar to anthrax and plague, with a false-positive rate of 0.4% |
| Model 3312A Ultraviolet Aerodynamic Particle Sizer (UV-APS) and Fluorescence Aerodynamic Particle Sizer-2 (FLAPS-2) | To detect living organisms in aerosols and nonvolatile liquids. | FLAPS-2 was able to detect 39 of 40 blind releases of stimulant aerosols (of particle ranging in size from 0.5 to 15 μm) at a distance of about 1 km with no false alarms during a 3-week period. In another trial, it was able to detect as few as 10 agent-containing particles per liter of air |
| Sensitive Membrane Antigen Rapid Test (SMART) and the Antibody-based Lateral Flow Economical Recognition Ticket (ALERT) | A handheld antigen/antibody test for the rapid detection of bioterrorism agents. | When field tested during the Gulf War, the SMART system had an “alarmingly” high false-positive rate thought secondary to contamination (14). SMART tests are reported per the manufacturer to have a |
aPCR; polymerase chain reaction. bWhere possible, we report sensitivity and specificity data (and highlight them in bold); if the published reports did not provide these values directly but did provide sufficient data for them to be calculated, we performed these calculations. cDenotes systems for which available evaluation data were from manufacturers’ Web sites only.
Evaluation data for diagnostic decision support systems for bioterrorism-related illnessa
| System name | Purpose | Evaluation datab |
|---|---|---|
| Clinical decision support system for detection and respiratory isolation of tuberculosis patients | To automate the detection and respiratory isolation of patients with positive cultures and chest x-rays suspicious for TB. | In a retrospective analysis, the system increased the proportion of appropriate TB isolations in inpatients from 51% to 75% but falsely recommended isolation of 27 of 171 patients. In a prospective analysis, the system correctly identified 30 of 43 of patients with TB but not identify 21 of these patients (false-negatives). However, the decision support system identified 4 patients not identified by the clinicians |
| Columbia–Presbyterian Medical Center Natural Language Processor | To automate the identification of 6 pulmonary diseases (including pneumonia) through analysis of radiology reports. | The system had a |
| Computer Program for Diagnosing and Teaching Geographic Medicine | To provide a differential diagnosis of infectious diseases matched to 22 clinical parameters for a patient; also to provide general information about infectious diseases, anti-infective agents, and vaccines. | The computer program correctly identified 75% (222 of 295) of the actual cases and 64% (128 of 200) of the hypothetical cases of patients with infectious diseases |
| DERMIS | To provide a differential diagnosis of skin lesions. | The system correctly diagnosed lesions 51% to 80% of the time and included the correct diagnosis among its top 3 choices 70% to 95% of the time (out of a total of 5,203 cases) |
| Dxplain | To provide a differential diagnosis based on clinician-entered signs and symptoms. The system includes descriptions and findings for potential bioterrorism agents, and is updated weekly to account for potential outbreaks. | In an evaluation of 103 consecutive internal medicine cases, Dxplain correctly identified the diagnosis in 73% of cases, with an average rank of 10.7 (the rank of a diagnosis refers to its position on the differential diagnosis—for example, the diagnosis with the greatest likelihood of being the actual disease is ranked first and the next most likely diagnosis is ranked second) |
| Fuzzy logic program to predict source of bacterial infection | To use age, blood type, gender, and race to predict the cause of bacterial infections. | The program was able to correctly classify 27 of 32 patients into 1 of 4 groups based on demographic data alone |
| Global Infectious Disease and Epidemiology Network (GIDEON) | To provide differential diagnoses for patients with diseases of infectious etiology. All potential bioterrorism agents as specified by CDC are included in the GIDEON knowledge base | Whereas medical house officers listed the correct diagnosis first in their admission note 87% of the time (for 75 of 86 patients), GIDEON provided the correct diagnosis for 33% (28 of 86 patients) |
| Iliad (and Medical HouseCall which is a system for consumers derived from Iliad) | To provide a differential diagnosis based on clinician-entered signs and symptoms. The knowledge base is focused in internal medicine and was last updated in 1997. | In a multicenter evaluation, each of 33 users analyzed 9 diagnostically difficult cases. On average, Iliad included the correct diagnosis in its list of possible diagnoses for 4 of the 9 cases, and included the correct diagnosis within its top 6 diagnoses for 2 of the 9 cases. The differential diagnosis generated by Iliad is not dependent upon the level of training of the user |
| Neural Network for Diagnosing Tuberculosis | To predict active pulmonary TB (using clinical and radiographic information) so that patients may be appropriately isolated at the time of admission. | The neural network correctly identified 11 of 11 patients with active TB ( |
| PNEUMON-IA | To diagnose community-acquired pneumonia from clinical, radiologic and laboratory data. | The decision support system correctly identified pneumonia in 4 of 10 cases, compared with between 3 and 6 cases for the clinician experts |
| Quick Medical Reference (QMR) | To provide a differential diagnosis based on clinician-entered signs and symptoms. | One prospective study used QMR to assist in the management of 31 patients for which the anticipated diagnoses were known to exist in the QMR knowledge base. In the 20 cases for which a diagnosis was ultimately made, QMR included the correct diagnosis in its differential in 17 cases (85%) and listed the correct diagnosis as most likely in 12 cases (60%) |
| SymText | To analyze radiology reports for specific clinical concepts such as identifying patients with pneumonia. | Average |
| Texas Infectious Disease Diagnostic Decision Support System | To provide a weighted differential diagnosis based on manually entered patient information. | The system was compared to a reference standard that missed the diagnosis of 98 of 342 cases of brucellosis. In 86 of the 98 patients, this system listed brucellosis in the top 5 diagnoses on the differential diagnosis list, and in 69 of these 98 patients, brucellosis was the only disease suggested by the system. The system missed the diagnosis in 12 of 98 patients. On average, without the system it took 17.9 days versus 4.5 days with the system to suspect the correct diagnosis |
| University of Chicago – Artificial Neural Network for Interstitial Lung Disease | To help radiologists differentiate among 11 interstitial lung diseases by using clinical parameters and radiographic findings to develop a differential diagnosis. | Areas under the ROC curve obtained with and without the system output were 0.911 and 0.826 (p < 0.0001), respectively |
| University of Chicago – Computer Aided Diagnosis of Interstitial Lung Disease | To aid in the detection of interstitial lung disease in digitized chest radiographs. | Areas under the ROC curve obtained with and without computer-aided diagnostic output were 0.970 and 0.948 (p |
aTB, tuberculosis; CDC, Centers for Disease Control and Prevention; ROC, receiver-operating characteristic curve. bWhere possible, we report sensitivity and specificity data (and highlight them in bold); if the published reports did not provide these values directly but did provide sufficient data for them to be calculated, we performed these calculations.
Figure 1Effect of sensitivity, specificity, and pretest probability on posttest probability of anthrax’s being present. Upper curves show the posttest probability of anthrax’s being present after a positive detection or diagnostic test result. Lower curves show the posttest probability of anthrax’s being present after a negative detection or diagnostic test result. Separate curves are drawn for two diagnostic tests described in the text: one with 99% sensitivity and 99% specificity (thick) and another with 96% sensitivity and 94% specificity (thin). The arrow marks a pretest probability of disease of 0.0014, which relates to the example described in the text.
Figure 2Receiver-operating characteristic curves (ROC). Each point along a ROC represents the trade-off in sensitivity and specificity, depending on the threshold for an abnormal test. Here, two hypothetical diagnostic tests are compared. The diagnostic test represented by the unbroken ROC curve is a better test than that represented by the broken ROC curve, as demonstrated by its greater sensitivity for any given specificity (and thus, greater area under the curve).