| Literature DB >> 29907560 |
Antony Hardjojo1, Arunan Gunachandran1, Long Pang1, Mohammed Ridzwan Bin Abdullah1, Win Wah1, Joash Wen Chen Chong1, Ee Hui Goh1, Sok Huang Teo2, Gilbert Lim3, Mong Li Lee3, Wynne Hsu3, Vernon Lee1, Mark I-Cheng Chen1,4, Franco Wong2,5, Jonathan Siung King Phang2,5.
Abstract
BACKGROUND: Free-text clinical records provide a source of information that complements traditional disease surveillance. To electronically harness these records, they need to be transformed into codified fields by natural language processing algorithms.Entities:
Keywords: communicable diseases; electronic health records; epidemiology; natural language processing; surveillance; syndromic surveillance
Year: 2018 PMID: 29907560 PMCID: PMC6026305 DOI: 10.2196/medinform.8204
Source DB: PubMed Journal: JMIR Med Inform
Figure 1Ontology and grammar-based analysis of the rule-based natural language processing (NLP) algorithm. Signs and symptoms and information on assertion status and duration are captured and tokenized in the ontology analysis. Relationships between tokens are built up in the grammar-based analysis. C/o: complain of; ST; sore throat.
Figure 2Sample set of clinical notes and transformation following phrase-level manual coding and episode-level coding. Abd: abdominal; NA: not applicable; NKDA: no known drug allergy; PMHX: past medical history; RIF: right iliac fossa.
Figure 3Flowchart of process for creating reference standard.
Frequency of the 10 most commonly detected signs and symptoms within 1680 primary care clinical records by human coders.
| Symptoms sorted by frequency of symptom mention in episode level | All instances (N=7703), n (%)a | Instance of symptom affirmation, n (%)b |
| Fever | 928 (12.04) | 228 (24.6) |
| Cough | 744 (9.66) | 646 (86.8) |
| Sore throat | 591 (7.67) | 416 (70.4) |
| Rhinorrhea | 552 (7.17) | 435 (78.8) |
| Altered state of consciousness | 376 (4.88) | 7 (1.9) |
| Vomiting | 347 (4.50) | 75 (21.6) |
| Rash | 345 (4.48) | 72 (20.7) |
| Dyspnea | 286 (3.71) | 31 (10.8) |
| Diarrhea | 271 (3.52) | 137 (50.6) |
| Sputum | 256 (3.32) | 212 (82.8) |
aColumn percentages, with the denominator being all instances (N=7703).
bRow percentages, with the denominator being the instances where the symptom in that row appears (eg, for Fever, n=928).
Phrase level precision, recall, and F-measure of human coders and Clinical History Extractor for Syndromic Surveillance (CHESS) outputs compared against instances of symptom occurrences in reference standard.
| Performance against reference standard | Comparison of coder 1 versus coder 2 (N=8861 instances) | CHESS performance for training set (after training of dictionary, N=4282 instances) | CHESS performance for validation set (after training of dictionary, N=4578 instances) | |
| Coder 1 | Coder 2 | |||
| Precision, % | 98.52 | 96.06 | 95.24 | 94.15 |
| Recall, % | 96.93 | 84.30 | 96.17 | 90.39 |
| F-measure, % | 97.72 | 89.80 | 95.70 | 92.23 |
aCHESS: Clinical History Extractor for Syndromic Surveillance.
Episode level precision, recall, and F-measure of human coders and Clinical History Extractor for Syndromic Surveillance (CHESS) outputs compared against instances of symptom occurrences in reference standard.
| Performance against reference standard | Comparison of coder 1 versus coder 2 (N=7703 instances) | CHESSa performance for training set (after training of dictionary, N=3738 instances) | CHESS performance for validation set (after training of dictionary, N=3965 instances) | |
| Coder 1 | Coder 2 | |||
| Precision, % | 98.91 | 97.13 | 96.74 | 95.97 |
| Recall, % | 97.46 | 88.47 | 97.65 | 93.06 |
| F-measure, % | 98.18 | 92.58 | 97.19 | 94.49 |
Figure 4Bubble chart of the Clinical History Extractor for Syndromic Surveillance’s (CHESS’s) precision and recall for each sign and symptom in episode level analysis for the validation dataset. Each bubble denotes a single symptom categorized into symptom types: respiratory, gastrointestinal, constitutional, and others. Bubble size is proportional to the number of cases identified by humans (true positive + false negative). Symptoms present in less than 1% of records are not presented.
Figure 5Clinical History Extractor for Syndromic Surveillance’s (CHESS’s) accuracy in identifying assertion status of symptoms within episode level analysis based on the validation dataset.
Figure 6Episode level analysis on the distribution of symptom episode duration in instances detected by human coders (blue) among all the National Healthcare Group Polyclinics (NHGP) records and the distribution of durations detected by Clinical History Extractor for Syndromic Surveillance (CHESS; red) based on the validation dataset. Diamonds give the proportion of records where CHESS correctly identifies and assigns the duration information stratified by episode duration (based on the reference standard), with the horizontal line giving the aggregated accuracy for detection of symptom duration for all records analyzed.