Literature DB >> 32574362

Can Unified Medical Language System-based semantic representation improve automated identification of patient safety incident reports by type and severity?

Ying Wang, Enrico Coiera, Farah Magrabi.   

Abstract

OBJECTIVE: The study sought to evaluate the feasibility of using Unified Medical Language System (UMLS) semantic features for automated identification of reports about patient safety incidents by type and severity.
MATERIALS AND METHODS: Binary support vector machine (SVM) classifier ensembles were trained and validated using balanced datasets of critical incident report texts (n_type = 2860, n_severity = 1160) collected from a state-wide reporting system. Generalizability was evaluated on different and independent hospital-level reporting system. Concepts were extracted from report narratives using the UMLS Metathesaurus, and their relevance and frequency were used as semantic features. Performance was evaluated by F-score, Hamming loss, and exact match score and was compared with SVM ensembles using bag-of-words (BOW) features on 3 testing datasets (type/severity: n_benchmark = 286/116, n_original = 444/4837, n_independent =6000/5950).
RESULTS: SVMs using semantic features met or outperformed those based on BOW features to identify 10 different incident types (F-score [semantics/BOW]: benchmark = 82.6%/69.4%; original = 77.9%/68.8%; independent = 78.0%/67.4%) and extreme-risk events (F-score [semantics/BOW]: benchmark = 87.3%/87.3%; original = 25.5%/19.8%; independent = 49.6%/52.7%). For incident type, the exact match score for semantic classifiers was consistently higher than BOW across all test datasets (exact match [semantics/BOW]: benchmark = 48.9%/39.9%; original = 57.9%/44.4%; independent = 59.5%/34.9%). DISCUSSION: BOW representations are not ideal for the automated identification of incident reports because they do not account for text semantics. UMLS semantic representations are likely to better capture information in report narratives, and thus may explain their superior performance.
CONCLUSIONS: UMLS-based semantic classifiers were effective in identifying incidents by type and extreme-risk events, providing better generalizability than classifiers using BOW.
© The Author(s) 2020. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For permissions, please email: journals.permissions@oup.com.

Entities:  

Keywords:  UMLS; incident reporting; natural language processing; patient safety; semantics; supervised machine learning; text classification

Mesh:

Year:  2020        PMID: 32574362      PMCID: PMC7566533          DOI: 10.1093/jamia/ocaa082

Source DB:  PubMed          Journal:  J Am Med Inform Assoc        ISSN: 1067-5027            Impact factor:   4.497


INTRODUCTION

Health systems are experiencing a growing need to better harness their incident monitoring systems to improve patient safety. Incident monitoring is now widespread with the implementation of monitoring systems in most developed nations. Healthcare professionals are routinely submitting reports about incidents or safety events in the delivery of care that could have resulted or did result in unnecessary harm to patients. By retrospectively analyzing incident reports, health systems are striving to detect and manage emerging risks to patient safety. The timely detection and response to patient safety incidents is highly dependent on rapid analysis of reports by humans, which is a highly resource-intensive task. Patient safety experts typically triage incidents by their type and severity level for detailed analyses of unstructured free text data to identify contributing factors so that lessons can be learned and corrective action can be taken. Incident types are usually based on priority areas for safety and quality improvement (eg, falls, medications), whereas severity level is based on the consequences of incidents and the likelihood of recurrence., For instance, hospital-level analyses could investigate common causes, contributing factors, and outcomes for specific incident types (eg, medications). Similarly, clusters of similar high-risk incidents could be identified at a health system level, indicating an emerging risk to patient safety. Take, for instance, an update to an order entry system that turns off the alerts for a high-risk medication across multiple sites. An incident at one site involving the medication may not be significant, whereas a cluster across multiple sites can facilitate early detection of the issue. A major limitation of manual analysis is that it can no longer keep up with the growing volume of reports. For instance, 203 140 incidents were reported in the Australian state of New South Wales in 2018. One way to improve the identification of incident clusters is to ask reporters to categorize incidents in a standardized manner using structures like the Agency for Healthcare Research and Quality Common Formats. However, the labeling of incidents only works if reporters are knowledgeable about classification systems and are able to apply them consistently. Studies have shown that the ratings provided by reporters are often inaccurate, and there is a high discordance in labels because health professionals may not have expertise in categorizing incidents. Moreover, labeling may often be absent, incomplete, or delayed, reducing the ability to respond in near real time. Another—potentially more effective—way to improve the efficiency of incident analyses is to apply machine learning methods. Recent work in this area has demonstrated the feasibility of using supervised text classification methods to automatically sift through the large volumes of incident reporting data and identify specific clusters of reports for further detailed analysis by humans. However, most studies have focused on using supervised methods for binary classification to identify a specific incident type. In reality, there are multiple incident types reflecting the breadth of problems in patient safety and little is known about the most optimal way to represent the free text from incident reports for automated identification. We recently demonstrated that a convolutional neural network with word embedding was effective in identifying incidents by type and severity, providing better generalizability than support vector machine (SVM) classifiers. While a convolutional neural network is an elegant solution to distinguish multiple incident types, it does not support multiple labels. Often an incident report can relate to more than 1 patient safety problem (ie, it can be assigned to multiple incident types). For example, “Episode label A for patient X was placed incorrectly onto the specimen that belongs to patient Y,” describes an error in patient identification that also relates to documentation. In another previous study to identify up to 2 labels for 10 incident types, we found that classifiers based on the bag-of-words (BOW) model failed when causes and consequences of incidents were implicitly described in reports. This is because BOW does not take the semantic structure of language into consideration, and the ordering of words is also lost. For instance, an incident about a deteriorating patient was misidentified as a medications problem because there was a long list of medications and their doses in the incident narrative. We also found that BOW classifiers were not effective in identifying rare classes of incidents (ie, making up < 2% of all incidents) in which the data available for training were limited (eg, infection, deteriorating patient). Thus, in this study, we sought to evaluate the utility of semantic feature representation for automated identification incident reports by type and severity. We chose the Unified Medical Language System (UMLS) Metathesaurus, as it covers more than 2.9 million medical concepts from more than 150 source biomedical vocabularies. The UMLS links alternative names for concepts and identifies useful relationships between concepts. The semantics in reports were represented by mapping incident narratives to UMLS concepts to train and validate (SVM) classifiers. For incident type, ensembles of binary classifier chains (ECCs) were used because they perform better than other multilabel classification algorithms, such as binary relevance,, and their chain structure can leverage the relationships among labels. We then compared the performance of this approach with BOW features from our previous work.,

MATERIALS AND METHODS

An overview of our approach comprising 4 main steps is shown in Figure 1.
Figure 1.

Flowchart to train, validate and test semantic classifiers using 3 testing datasets: benchmark, original, and independent.

Flowchart to train, validate and test semantic classifiers using 3 testing datasets: benchmark, original, and independent. Data preprocessing: Incident reports were collected from 2 separate reporting systems, the Advanced Incident Management System (AIMS) and Riskman., The reports were labeled by patient safety experts to provide a gold standard label for experiments. Semantic features representation: The narratives of reports were processed to generate semantic features. Training and validating classifiers: Classifiers were trained and validated under cross-validation process to select optimal models based on performance. Here, the multilabel classification problem was transformed by training multiple basic binary classifier chains (CCs) (ie, developing a classifier identifying one type against all others while involving the relationships among incident types). For severity level, we decomposed the problem into a series of binary classification problems. To distinguish 4 severity levels, 6 binary classifiers were trained using balanced datasets with one-vs-one (OvsO) strategy. Testing classifiers: Was undertaken with 3 testing datasets from the AIMS and Riskman systems. Incident type and severity level were examined separately. Model performance was evaluated and compared with BOW classifiers. Each of these 4 steps is further detailed subsequently.

Step 1: Data preprocessing

Incident reporting systems

The reports were collected from 2 separate incident reporting systems: AIMS and Riskman., AIMS has been used in Australia, New Zealand, South Africa, and the United States. In Australia, it has been used across the public hospital system in 4 of the 8 states and territories: New South Wales, Western Australia, South Australia, and the Northern Territory. In one Australian state, 137 522 incidents were reported to AIMS from January to December 2011. Of these, 6000 reports were selected using a random sampling approach. Similarly, an independent set of 6000 reports were randomly selected from those that had been submitted to a hospital-level Riskman system from January 2005 to July 2012. The Riskman system is an independent tool used across the state of Victoria and a number of private hospitals. Incident reports consist of a number of structured and free text fields to describe the event and its consequences (Box 1). Upon collection, any personally identifiable information was removed in accordance with jurisdictional privacy requirements (eg, name, date of birth). Only descriptive narratives were used in experiments including incident description, patient outcome, actions taken, prevention steps, investigation findings and results. All system-specific codes, punctuation, and nonalphanumerical characters were removed, and text was converted to lowercase. Ethical approval was obtained from university committees as well as a committee governing the hospital and state datasets. incident type(s) severity access code preventative steps patient outcome

Labeling reports

Given the inconsistency of reporter labels, 3 experts in the classification of incidents reviewed and validated labels for all reports based on the international classification for patient safety. The reports were split among the 3 experts, and the labels they provided were used as “gold standard” for training and testing the performance of classifiers. The reports were classified into 20 incident types, and this study focused on 10 of these types that have been recognized as priority areas for safety and quality improvement (Table 1).,, Interrater reliability for determining incident types was Cohen’s kappa = 0.93 (P < .001; 95% confidence interval, 0.9301-0.9319). To cover the whole dataset, an “Others” set was created using a random sampling approach to ensure representativeness of 10 other unrelated incident types—see Appendix B in in Wang et al.
Table 1.

Composition of balanced and stratified datasets for training and testing semantic classifiers

Benchmark (balanced AIMS)
Original (stratified AIMS)
Independent (stratified Riskman)
N1N2N1N2%N1N2%
Incident type
 Falls26026190912087293915
 Medications2603046874151053121718
 Pressure injury260264373881901973
 Aggression2602714957114875418
 Documentation260589266762528094
 Blood producta26027356159701
 Patient identificationa260337782861171
 Infectiona2602746612235<1
 Clinical handovera260301782871011
 Deteriorating patienta26026412<11421<1
 Others260689148173332,878403948
Total2860382744453060008086
Severity level
 SAC1

290

290

290

290

1160

25

95

2198

2519

4837

<1

23

105

2609

3213

5950

<1
 SAC222
 SAC34544
 SAC45254
Total

The same data was used to train bag-of-words classifiers.

N1 is the number of reports based on primary labels. N2 is the number of reports considering 2 labels, and % is based on primary label alone.

AIMS: Advanced Incident Management System; SAC: severity assessment code.

Rare incident type (ie, <2%).

Composition of balanced and stratified datasets for training and testing semantic classifiers 290 290 290 290 1160 25 95 2198 2519 4837 23 105 2609 3213 5950 The same data was used to train bag-of-words classifiers. N1 is the number of reports based on primary labels. N2 is the number of reports considering 2 labels, and % is based on primary label alone. AIMS: Advanced Incident Management System; SAC: severity assessment code. Rare incident type (ie, <2%). The seriousness of an incident was rated by an internationally accepted rating system called the severity assessment codes (SACs), developed by the U.S. Veterans Administration. Given the severity of an incident and the likelihood of recurrence, 4 risk ratings (i, extreme; ii, high; iii, medium; and iv, low) were used. The gold standard was based on assignment of SAC ratings for every report by local patient safety managers who had received training in assessing incident seriousness and were familiar with the nature of incidents and their consequences.

Generating training and testing subsets

Of the 6000 AIMS reports, we used 260 that were randomly selected from each incident type and 290, also randomly selected, for each SAC level to create a balanced dataset (Table 1). The sample sizes were based on previous studies. The balanced dataset was further divided into training (80%), validation (10%), and testing (10%) subsets under a 10-fold subsampling cross-validation process. The training and validation subsets were used to identify the most effective classifiers, and the testing subset (benchmark) was applied to generate benchmark results. To evaluate applicability in real-world conditions, classifiers were further tested on imbalanced (ie, “stratified”) datasets from AIMS (original). These stratified datasets were randomly selected from the remaining AIMS reports based on the real-world ratio incidents by type and severity (Table 1). To examine generalizability to an independent incident reporting system, classifiers were tested on a stratified Riskman dataset (independent).

Step 2: Semantic features representation

We developed a bag-of-concepts model to extract unique concepts as semantic features from incident narratives. This approach is similar to a BOW model extracting unique words but maps tokens, terms, and words from incident reports into medical concepts from the UMLS. To identify unique concepts, a software tool called MetaMap was used to annotate the UMLS concepts and semantic types. The bag-of-concepts model was represented as an unordered collection of concepts in which each concept was used as a feature. Three feature engineering approaches were adopted to represent the occurrence of concepts: binary count, term frequency (TF), and TF–inverse document frequency (TF-IDF).

Step 3: Training and validating classifiers

Ensemble strategy

The classification problem was decomposed into a number of binary classification problems., There are 2 traditional ensemble strategies to pool decisions from each base binary classifier: one-vs-all (OvsA) and OvsO. OvsA divides an n class problem into n binary problems by training classifiers to distinguish one class from all other classes., It was used in identifying incident types as it considered all possible label combinations, suiting the design of learning label relationships., The OvsO strategy transforms an n class problem into n*(n-1)/2 binary problems by involving all possible combinations between pairs of classes. Given its good performance to identify severity levels in our previous work, the OvsO strategy was adopted.

Base binary classifiers

SVM with radial basis function (RBF) kernel was chosen, as it has been shown to be effective for smaller datasets with a large feature space., Furthermore, SVM outperformed other binary classifiers in our previous work, such as a logistic regression model., We used an ensemble of SVM classifiers with semantic features., For incident type, SVM classifiers were linked along a chain where each classifier dealt with a binary problem. Classifier chains were different from the common binary classifiers because their feature space was extended with binary values to indicate the relationships between incident types that appeared in training data. The order of the chain itself was determined by the order of the label variables. In the extended feature space, binary values of labels only indicated which of previous labels were assigned to reports. To reduce the influence of label order, incident types were randomly reordered and training was repeated. Final decisions were based on the predictions of each ECC. As classification performance often increases with ECC size used in training, ECCs may become unnecessarily large and impose a high computational cost. We examined the optimal size for ECC combinations by varying the number of ensembles from 3 to 40. A learning curve was used to find the optimal ECC by trading off computational cost against classifier accuracy. The feature space for severity level involved the sematic features alone. This is because severity level involves a single label or risk group while there may be multiple incident types. Base classifiers were trained, validated, and tested on balanced datasets using 3 semantic feature representations: binary count, TF, and TF-IDF. A 10-fold subsampling cross-validation method was applied to optimize the classifier parameters. The parameters from classifiers achieving the highest F-score were adopted for testing.

Step 4: Testing classifiers

In group decision making, majority vote is an efficient method and has been commonly accepted in classification ensembles. To identify incident type, the final decisions were made by averaging multiple predictions from individual CCs. Up to 2 labels were predicted if the averaged classification probabilities exceeded a predefined threshold. For severity level, each base classifier voted and the final prediction was based on the level with the most votes. For incident type, overall performance was evaluated using example-based and label-based measures. Example-based measures evaluate the difference of the true and the predicted sets of labels over all testing reports. Label-based measures consider additional degrees of freedom with multilabels and evaluate performance separately for each incident type, and then average the performance over all types. We used 6 commonly example-based measures, including Hamming loss, accuracy, exact match score, precision, recall, and F-score. Given 2 labels per report, the performance for individual types was evaluated based on OR logic when matching the predicted and true set of labels. Label-based measures included macro-averaged and micro-averaged precision, recall, and F-score., We also examined the top 20 UMLS concepts from each incident type to identify concepts that were common across the 10 types of incidents, as well as those that were unique to each incident type. For severity level, overall performance was evaluated using micro-averaged measures of precision, recall, and F-score based on the cumulative number of true positives, true negatives, false positives, and false negatives per type., When identifying a specific severity level, the F-score, precision, and recall measures were evaluated per level. Individual measures were calculated based on the probability that a specific severity level was classified as such (eg, % of extreme-risk incidents correctly identified among the test set for SAC1).

RESULTS

Overall performance to identify incident type

For incident type, the most effective classifier was an ensemble of 12 binary classifier chains of SVM RBF kernel with TF-IDF UMLS-based semantic feature representation, achieving an averaged F-score of 82.6% on the benchmark, 77.9% on the original, and 78.0%, on the independent datasets (Table 2).
Table 2.

The most effective classifiers for incident type using semantic features compared with BOW

Classification studiesSemanticsBOW15
Ensemble strategyOvsAOvsA
Ensemble size12 ECCs6 ECCs
Feature extractionBag of conceptsBOW
Feature space representationTF-IDF + label associationsBinary count + label associations
Base classifierSVM RBF kernelSVM RBF kernel
Group decision makingVotingVoting
Average F-score
Benchmark dataset, %82.669.4
Original dataset, %77.968.8
Independent dataset, %78.067.4

BOW: bag-of-words; ECCs: ensemble of binary classifier chains; OvsA: one vs all; RBF: radial basis function; SVM: support vector machine; TF-IDF: term frequency–inverse document frequency.

The most effective classifiers for incident type using semantic features compared with BOW BOW: bag-of-words; ECCs: ensemble of binary classifier chains; OvsA: one vs all; RBF: radial basis function; SVM: support vector machine; TF-IDF: term frequency–inverse document frequency. Compared with BOW, the overall performance of semantic classifiers was superior (Table 3). When identifying multiple incident types to which reports could be assigned, semantic classifiers achieved better exact match scores, improving overall performance on all 3 testing datasets. Hamming loss dramatically dropped from 7.2 to 3.7 on the original dataset and decreased from 8.1 to 3.9 on the independent dataset. By label-based measures, both semantic and BOW classifiers performed consistently on the original and independent datasets. With the exception of macro-recall, all label-based measures were higher.
Table 3.

Overall classification performance of semantic features in identifying incident type compared with BOW

Benchmark
Original
Independent
Feature representationSemanticsBOWSemanticsBOWSemanticsBOW
Example-based measures
 Hamming loss3.97.83.77.23.98.1
 Accuracy82.264.475.668.075.461.7
 Precision84.770.677.372.978.166.4
 Recall84.477.176.776.676.867.4
 F-score84.673.777.074.777.466.9
 Exact match48.939.957.944.459.534.9
Label-based measures
 Macro-precision86.769.777.752.470.254.2
 Macro-recall91.179.079.977.478.270.3
 Macro–F-score87.973.777.159.273.658.8
 Micro-precision82.867.178.067.178.968.7
 Micro-recall82.471.977.970.777.166.1
 Micro–F-score82.669.477.968.878.067.4

Values are %.

BOW: bag-of-words.

Overall classification performance of semantic features in identifying incident type compared with BOW Values are %. BOW: bag-of-words.

Identifying individual incident types

On the benchmark dataset, semantic classifiers outperformed BOW when identifying all incident types except falls and deteriorating patient, which remained steady at 88.1% and 87.5%, respectively (Table 4). With the original dataset, F-score improved, ranging from 64% for patient identification to 92.4% for falls. On the independent dataset, F-scores improved for all incident types except documentation, which remained steady at 75.2% (BOW: 75%), and for deteriorating patient, in which it worsened to 48.8% (BOW: 54.6%).
Table 4.

F-score for identifying incident types and severity level using semantic features compared with BOW

Benchmark
Original
Independent
Feature representationSemanticsBOW12,15SemanticsBOW12,15SemanticsBOW12,15
Incident type
 Falls88.188.192.489.790.381.0
 Medications78.567.487.176.385.175.2
 Pressure injury96.491.591.685.485.781.4
 Aggression93.174.082.669.179.363.8
 Documentation71.062.278.261.275.275.0
 Blood productsa96.671.175.044.476.954.9
 Patient identificationa77.266.064.034.070.556.4
 Infectiona93.881.180.046.263.130.8
 Clinical handovera77.062.551.632.050.434.5
 Deteriorating patienta87.589.766.744.448.854.6
 Others75.860.078.667.983.469.4
Severity level
 Average F-score71.662.942.250.149.652.7
 SAC187.387.325.519.850.712.5
 SAC269.849.08.412.311.912.0
 SAC375.349.142.142.658.448.3
 SAC460.064.052.261.839.360.0

Values are %.

BOW: bag-of-words; SAC: severity assessment code.

Rare incident type (ie, <2%).

F-score for identifying incident types and severity level using semantic features compared with BOW Values are %. BOW: bag-of-words; SAC: severity assessment code. Rare incident type (ie, <2%). When identifying the rare incident types of blood products, patient identification, infection, and clinical handover, performance of the semantic classifiers improved dramatically on the original and independent datasets (Table 4). For instance, the F-score for blood products increased from 44.4% to 75% on the original; and from 54.9% to 76.9% on the independent. Patient identification incidents were frequently misclassified by BOW as documentation and others (F-score: original = 34.0%; independent = 56.4%). However, semantic classifiers improved this performance to 64.0% and 70.5%. For infection, F-scores were also improved (original: 46.2% to 80%; independent 30.8% to 63.1%). For clinical handover, F-scores for the semantic classifiers ranged from 50.4% to 51.6%, compared with from 32% to 34.5% with BOW. The exception was deteriorating patient, in which F-score improved on the original dataset (from 44.4% to 66.7%), but was worse on the independent dataset (from 54.6% to 48.8%).

Mapping between UMLS concepts and incident types

We reviewed the top 20 concepts from each incident type (see Supplementary Appendix). Not surprisingly, the UMLS concept that was common across all types was patients. Other concepts that were common across 8-9 incident types were ward, result, present, and time (see Supplementary Appendix Table 5). Physicians, nurses, and notification were common across 6 incident types. Concepts that were unique to individual incident types generally related to specific clinical tasks and procedures (see Supplementary Table 6). For instance, the concepts dosage, prescribed, pharmacy, and pharmacist infusion procedures were unique to medications, whereas secluding patient, aggressive behavior, and escort were unique to aggression incidents. For blood products, unique concepts included peripheral blood, blood transfusion, blood product, blood bank, and departments such as laboratory and pathology. For documentation, unique concepts included signature and electronic health records, whereas transfer of care and communication were unique to clinical handover.

Identifying severity level

For severity level, OvsO ensembles of SVM RBF kernel with TF-IDF feature representation was the most effective combination. While the micro-averaged F-score (71.6%) was higher than BOW on the benchmark, it was lower on the original (42.2%) and independent (49.6%) datasets (Table 4). For SAC1, performance of the semantic classifiers was consistent with BOW on the benchmark (F-score = 87.3%) but markedly improved on the original (F-score up from 19.8% to 25.5%) and independent (F-score up from 12.5% to 50.7%) datasets. For SAC2 and SAC3, semantic classifiers performed better than BOW on the benchmark dataset (SAC2: F-score up from 49.0% to 69.8%; SAC3: F-score up from 49.1% to 75.3.8%), but this improvement did not carry over to the original and independent datasets. For SAC4, F-score was consistently lower than BOW across all testing datasets.

DISCUSSION

Main findings and implications

We evaluated the feasibility of a UMLS-based semantic feature representation by comparing it with BOW features and found that it can enhance the performance of SVM classifiers in identifying patient safety incidents by type and severity, particularly high-risk SAC1 events. To the best of our knowledge, no previous studies have compared these 2 approaches to represent the free text from incident reports for automated identification. In terms of overall performance, semantic feature representation improved generalizability, showing consistent performance from the original (F-score: 77.9%) to independent (F-score: 78.0%) datasets. Identification of primary and secondary labels was also improved; semantic classifiers achieved exact match scores of 57.9% and 59.5% on the original and independent datasets (BOW: 44.4%/34.9%). For the different incident types, semantic classifiers improved identification of 8 types (ie, medications, pressure injury, aggression, documentation, blood products, patient identification, infection, and clinical handover). F-score was above 75% for 7 of 10 types in the original dataset. These 7 types made up 95% of all reported incidents (Table 1). There was a marked improvement in identifying rare incident types and extreme-risk events, indicating that UMLS concepts better capture information in report narratives compared with BOW and may thus be an effective strategy to address the highly unbalanced distribution of incident classes, particularly for rare incident types and high-risk incidents, which make up <2% of incidents. These results show that semantic classifiers have the potential to be applied in real-world settings. When human resources are lacking, automated methods can reduce effort by supporting the first step in incident analysis. Rapidly screening which incidents require immediate attention, and grouping incidents by type, allows human efforts to focus on a much smaller volume of incident reports. Automated methods are currently meant to be used in conjunction with expert review and are not intended to replace human input. While the feasibility of automated incident identification has been demonstrated in controlled experiments, further work is required to understand the extent to which these methods can be generalized and how best they can be adapted to new settings to help identify clusters of potentially related reports with an underlying common cause. There is a need to trial text classifiers in real-world settings and to shift research from its sole focus on algorithmic performance to studying, in parallel, strategies for successful implementation and impact of these models on quality improvement initiatives and patient safety outcomes. Any translation, sharing, and reuse of text classifiers should seek to leverage the growing use of common data formats and platforms for machine-executable models.

Feature representation

We noted that the most effective feature representation for semantic classifiers was TF-IDF and binary count for BOW. TF-IDF reflects how important a word is to a report from all training documents, and its term-weighting scheme decreases the weight for commonly used concepts and increases the weight for infrequent concepts (ie, those might be specific to individual rare types). This might explain why the semantic features represented more distinctive concepts from rare classes and achieved better performance. Many clinical tasks and procedures unique to specific incident types were captured quite well by the UMLS concepts (Supplementary Table 6), contributing to better performance. However, these concepts did not focus on patient safety. To improve classification performance, more broadly applicable concepts for specific incident types are required. Better UMLS coverage of essential patient safety terminologies would be helpful for large-scale analysis of patient safety incidents. Another approach for incident types not adequately covered is to combine UMLS concepts with BOW features. This is an area that requires further investigation. For severity level, semantic features improved the identification of SAC1 incidents in stratified datasets. On the independent dataset, F-score increased dramatically from 12.5% (BOW) to 50.7% (semantics). However, there was no consistent improvement for SAC2 and SAC3 incidents. SAC4 incidents were more likely misclassified as SAC3 and SAC2 by the semantic classifiers, achieving a lower F-score. As SAC4 is a common class, the average F-score across all levels was lowered. SAC4 was mostly misclassified as SAC3. We observed that the reports from SAC3 and SAC4 covered all 10 incident types, making it hard for them to be distinguished from each other, as the training reports from both severity levels generated similar semantic feature spaces. In addition, the involvement of 10 incident types resulted in a sparse semantic feature space to cover all the concepts representing the different clinical tasks and procedures. In general, sparse input requires more training samples to achieve reliable prediction. However, we did not have access to a larger training dataset. One possible solution is to apply feature engineering, such as feature selection, to obtain more informative features but with lower dimensionality. The semantic feature spaces from the medium and low severity levels of SAC3 and SAC4 were overlapped. The incidents from these 2 levels covered similar types of incidents, leading to the harder boundary in feature space for distinguishing them.

Limitations and future work

First, we used datasets from 2 independent reporting systems, but both came from the same Australian state. Therefore, the semantic classifiers may not be generalizable to other jurisdictions and regions using different reporting, linguistic styles, and terminology. Second, the training datasets were balanced. Given the class imbalance problem with incident types in real-world settings, a stratified training dataset may be more desirable when large training datasets are available. Last, to identify the different incident types, we used the same threshold to make final decisions when evaluating results from binary classifiers. We observed that classifiers identified common types with more confident probabilities (eg, most falls achieved a probability higher than 0.9), but the probabilities for rare classes were relatively lower (eg, infection ∼0.5). This indicates that the same threshold for common and rare types may not be optimal, as it may result in some reports belonging to rare types being missed. Further experiments are required to investigate optimal thresholds for common and rare incident types.

CONCLUSION

Our experiments showed that semantic classifiers using UMLS concepts outperformed classifiers based on BOW features for identifying most incident types and extreme risk events. Semantic classifiers demonstrated better generalizability on the independent dataset by combing TF-IDF feature representation. Analysis of concepts unique to specific incident types indicated that semantic representation provides more distinct features to specify clinical tasks and procedures and may thus improve identification of incident types. Further work is required to improve identification of lower severity levels where decision boundaries are overlapped.

FUNDING

This research is supported in part by grants from the Australian National Health and Medical Research Council, project grant APP1022964 (FM); and the Centre for Research Excellence in Digital Health, grant 1134919 (EC and FM). The funding source did not play any role in the study design; collection, analysis, and interpretation of data; writing of the report; or decision to submit the article for publication.

AUTHOR CONTRIBUTIONS

EC, FM, and YW conceptualized the study. YW designed and implemented the training and evaluation of classification model, and is responsible for the integrity of the work. YW and FM drafted the article. All authors participated in writing and revising the article. All aspects of the study (including design; collection, analysis and interpretation of data; writing of the report; and decision to publish) were led by the authors. All authors read and approved the final manuscript.

ETHICS APPROVAL

Ethical approvals were obtained from committees of Macquarie University, the University of New South Wales as well as the committees governing the hospital and the state datasets.

SUPPLEMENTARY MATERIAL

Supplementary material is available at Journal of the American Medical Informatics Association online. Click here for additional data file.
Report formatStructuredFree text
Basic elementincident IDdescription of incident
date and timeactions taken

incident type(s)

severity access code

preventative steps

patient outcome

investigation findings and results
  17 in total

1.  The Unified Medical Language System (UMLS): integrating biomedical terminology.

Authors:  Olivier Bodenreider
Journal:  Nucleic Acids Res       Date:  2004-01-01       Impact factor: 16.971

2.  An overview of MetaMap: historical perspective and recent advances.

Authors:  Alan R Aronson; François-Michel Lang
Journal:  J Am Med Inform Assoc       Date:  2010 May-Jun       Impact factor: 4.497

3.  Automated Classification of Multi-Labeled Patient Safety Reports: A Shift from Quantity to Quality Measure.

Authors:  Chen Liang; Yang Gong
Journal:  Stud Health Technol Inform       Date:  2017

Review 4.  A systematic review of natural language processing for classification tasks in the field of incident reporting and adverse event analysis.

Authors:  Ian James Bruce Young; Saturnino Luz; Nazir Lone
Journal:  Int J Med Inform       Date:  2019-10-05       Impact factor: 4.046

Review 5.  Developing and deploying a patient safety program in a large health care delivery system: you can't fix what you don't know about.

Authors:  J P Bagian; C Lee; J Gosbee; J DeRosier; E Stalhandske; N Eldridge; R Williams; M Burkhardt
Journal:  Jt Comm J Qual Improv       Date:  2001-10

6.  Inconsistency in classification and reporting of in-hospital falls.

Authors:  Terry P Haines; Bernadette Massey; Paul Varghese; Jennifer Fleming; Len Gray
Journal:  J Am Geriatr Soc       Date:  2009-01-29       Impact factor: 5.562

7.  Using multiclass classification to automate the identification of patient safety incident reports by type and severity.

Authors:  Ying Wang; Enrico Coiera; William Runciman; Farah Magrabi
Journal:  BMC Med Inform Decis Mak       Date:  2017-06-12       Impact factor: 2.796

8.  The Last Mile: Where Artificial Intelligence Meets Reality.

Authors:  Enrico Coiera
Journal:  J Med Internet Res       Date:  2019-11-08       Impact factor: 5.428

9.  Computable knowledge: An imperative for Learning Health Systems.

Authors:  Charles P Friedman; Allen J Flynn
Journal:  Learn Health Syst       Date:  2019-10-06

10.  Towards an International Classification for Patient Safety: key concepts and terms.

Authors:  William Runciman; Peter Hibbert; Richard Thomson; Tjerk Van Der Schaaf; Heather Sherman; Pierre Lewalle
Journal:  Int J Qual Health Care       Date:  2009-02       Impact factor: 2.038

View more
  1 in total

1.  The UMLS knowledge sources at 30: indispensable to current research and applications in biomedical informatics.

Authors:  Betsy L Humphreys; Guilherme Del Fiol; Hua Xu
Journal:  J Am Med Inform Assoc       Date:  2020-10-01       Impact factor: 4.497

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.