| Literature DB >> 33059690 |
Enrico Mensa1, Davide Colla1, Marco Dalmasso2, Marco Giustini3, Carlo Mamo2, Alessio Pitidis3,4, Daniele P Radicioni5.
Abstract
BACKGROUND: Emergency room reports pose specific challenges to natural language processing techniques. In this setting, violence episodes on women, elderly and children are often under-reported. Categorizing textual descriptions as containing violence-related injuries (V) vs. non-violence-related injuries (NV) is thus a relevant task to the ends of devising alerting mechanisms to track (and prevent) violence episodes.Entities:
Keywords: Categorization explanation; Event extraction; Explanation; Semantic frames; Slot filling; Text categorization; Violent event tracking; Word embeddings; XAI
Mesh:
Year: 2020 PMID: 33059690 PMCID: PMC7559980 DOI: 10.1186/s12911-020-01237-4
Source DB: PubMed Journal: BMC Med Inform Decis Mak ISSN: 1472-6947 Impact factor: 2.796
Fig. 1The system outline. A complete outline of the VIDES system. The medical records initially undergo a cleaning step, they are then categorized into violent and non violent ones; subsequently records deemed to contain violence-related injuries are selected for further processing, in order to obtain an explanation of such categorization
Fig. 2Distribution of out-of-vocabulary terms. Distribution of the 50 most frequent acronyms, abbreviations and misspelled terms in the dataset
Fig. 3The neural network architecture. The neural architecture employed for the categorization task
Compatibility table illustrating the allowed PoSs and SSTs for each explanation frame field
| Noun | Noun_Person | |
| Noun | Noun_Object | |
| Noun_Artifact | ||
| Noun_State | ||
| Noun_Substance | ||
| Noun_Feeling, Noun_Act | ||
| Noun | Noun_Time | |
| Noun | Noun_Location | |
| Noun | Noun_Body | |
| Noun, Adjective | Noun_State, Adj_all | |
| Noun_Phenomenon |
Fig. 4Example of frame extraction for a sentence. The sentence extracted from the medical record is initially preprocessed, and then given in input to the frame extraction process. Every word in the sentence is considered as a potential candidate for each role. Candidates are then filtered and ranked, and the top scoring one is selected in order to obtain the final filler for each frame element
Precision, Recall and F1 scores for violence (V) and non-violence (NV) classes on the test set
| midrule NV | .99 | 1.0 | .99 |
| V | .92 | .80 | .86 |
Results for the explanation algorithm along with the baseline
| Baseline | .12 | .12 | .14 | .13 | |
| .24 | .22 | .29 | .27 | ||
| .73 | .74 | .74 | .73 | ||
| .50 | .51 | .51 | .50 | ||
| .15 | .18 | .26 | .18 | ||
| .30 | .36 | .37 | .31 | ||
| Main algorithm | .58 | .59 | .60 | .58 | |
| .28 | .31 | .32 | .29 | ||
| .80 | .82 | .82 | .80 | ||
| .90 | .90 | .90 | .90 | ||
| .49 | .57 | .57 | .49 | ||
| .40 | .45 | .47 | .41 |
Precision, Recall and F1 scores for violence (V) and non-violence (NV) classes on the test set, after correction of the mistakenly annotated false positives
| NV | .99 | 1.0 | .99 |
| V | .97 | .81 | .88 |
Precision, Recall and F1 scores for violence (V) and non-violence (NV) classes on the test set, after correction of the mistakenly annotated false positives and deletion of false negatives
| NV | 1.0 | 1.0 | 1.0 |
| V | .97 | .94 | .95 |