| Literature DB >> 21658290 |
Rezarta Islamaj Doğan1, Aurélie Névéol, Zhiyong Lu.
Abstract
BACKGROUND: Patient records contain valuable information regarding explanation of diagnosis, progression of disease, prescription and/or effectiveness of treatment, and more. Automatic recognition of clinically important concepts and the identification of relationships between those concepts in patient records are preliminary steps for many important applications in medical informatics, ranging from quality of care to hypothesis generation.Entities:
Mesh:
Year: 2011 PMID: 21658290 PMCID: PMC3111589 DOI: 10.1186/1471-2105-12-S3-S3
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Figure 1The diagram of clinical relationships Concepts appear in blue boxes, while the relationships between them appear in coloured diamonds.
Examples of relationships between medical concepts in patient records.
| Relationship | Abbrev. | I2b2 Notation | Example Sentence |
|---|---|---|---|
| Treatment Improves Medical Problem | Improves | TrIP | |
| Treatment Worsens Medical Problem | Worsens | TrWP | He was started on |
| Treatment Causes Medical Problem | Causes | TrCP | During ER evaluation pt was noted to have some degree of |
| Treatment is Administered for Medical Problem | Given | TrAP | |
| Treatment is Not Administered because of Medical Problem | Not Given | TrNAP | There was some question regarding restarting of the patient 's |
| Test Reveals Medical Problem | Reveals | TeRP | |
| Test is Conducted to Investigate Medical Problem | Conducted | TeCP | There was some concern that the patient may have |
| Medical Problem Indicates Medical Problem | Relates | PIP |
* Medical problems are shown in italics, tests are shown in bold and treatments are underlined.
Figure 2Relationship representation between two concepts as five context-blocks Five context-blocks: introductory block—the set of words from the beginning of the sentence to the occurrence of the first concept, 1 block—the set of words that comprise first concept (not necessarily first in the sentence), connective block—the set of words that tie the two concepts in the relationship, 2 block—the set of words that comprise second concept, and conclusive block—the set of words from the 2nd concept to the end of the sentence.
Data description.
| Relationship | Number of Positive examples in training set | Number of positive examples in testing set |
|---|---|---|
| Improves | 107 | 198 |
| Worsens | 56 | 143 |
| Causes | 296 | 444 |
| Given | 1421 | 2486 |
| Not Given | 106 | 191 |
| Reveals | 1733 | 3033 |
| Conducted | 303 | 588 |
| Relates | 1239 | 1986 |
The aggregated number of examples extracted from 349 patient records in the training set, and 477 patient records in the test set.
Examples of medical concepts in patient records.
| Medical Concept | Example Sentence |
|---|---|
| Problem | On admission , the patient was found to have |
| Treatment | Infectious Disease was consulted and recommended |
| Test |
*Medical problems are shown in italics, tests are shown in bold and treatments are underlined.
Examples of assertion categories of problem concepts in patient records.
| Example Sentence | Problem | Assertion |
|---|---|---|
| On admission , the patient was found to have | “a mild fever” | Present |
| “Bartonella” | Possible | |
| It would be useful to follow this at outpatient if the patient is | “symptomatic” | Hypothetical |
| He denied ever having any | “chest pain” | Absent |
| Patient has a family history of | “coronary artery disease” | Associated with someone else |
Concept identification
| Concept | Exact span evaluation | Inexact span evaluation | ||||
| Precision | Recall | F-measure | Precision | Recall | F-measure | |
| Problem | 0.802 | 0.818 | 0.810 | 0.911 | 0.914 | 0.912 |
| Treatment | 0.929 | 0.891 | 0.909 | 0.975 | 0.946 | 0.960 |
| Test | 0.943 | 0.893 | 0.917 | 0.975 | 0.934 | 0.954 |
| Best i2b2 system | 0.869 | 0.836 | 0.852 | 0.932 | 0.917 | 0.924 |
Performance evaluation for the relates relationship, using string matching and SVM models.
| Relationship Model | Precision | Recall | F-measure |
|---|---|---|---|
| String Matching | 0.177 | 0.511 | 0.263 |
| Naïve bag-of-words SVM | 0.254 | 0.960 | 0.402 |
| Context-blocks SVM (words) | 0.601 | 0.796 | 0.685 |
| Context-blocks SVM (Words + Assertion) | 0.598 | 0.784 | 0.679 |
| Context-blocks SVM (Words + CUI) | 0.746 | ||
| Context-blocks SVM (Words + SemType) | 0.590 | 0.678 |
Performance evaluation for the best models of all relationships
| Improves | 0.489 | 0.619 | 0.607 | SemTyp | |
| Worsens | 0.481 | 0.800 | 0.358 | Assertion | |
| Causes | 0.470 | 0.644 | 0.588 | CUI, Assertion, SemTyp | |
| Given | 0.713 | 0.727 | 0.872 | CUI, Assertion, SemTyp | |
| Not Given | 0.618 | 0.800 | 0.604 | CUI | |
| Reveals | 0.772 | 0.805 | 0.932 | - | |
| Conducted | 0.533 | 0.584 | 0.742 | Assertion | |
| Relates | 0.543 | 0.646 | 0.746 | CUI | |
Per-relationship and per-record f-measures computed prior to and after feature selection.
| Feature Selection | Prior | After | Prior | After |
|---|---|---|---|---|
| Improves | 0.613 | 0.683 | 0.703 | 0.814 |
| Worsens | 0.494 | 0.673 | 0.569 | 0.621 |
| Causes | 0.615 | 0.735 | 0.695 | 0.793 |
| Given | 0.793 | 0.866 | 0.768 | 0.831 |
| Not Given | 0.688 | 0.754 | 0.685 | 0.743 |
| Reveals | 0.864 | 0.912 | 0.819 | 0.855 |
| Conducted | 0.654 | 0.759 | 0.803 | 0.883 |
| Relates | 0.692 | 0.775 | 0.761 | 0.823 |
Figure 3Comparison between features that represent relationship The sentence blocks are shown sequentially for conducted on the right, and for reveals on the left. Each sentence block is named, and the positively selected features are highlighted in the green block, while the negatively weighted features are highlighted in the red block. From this diagram, we can see that some features which are weighted highly positive for one relationship are weighted, in fact, negatively for the other.
F-measures computed prior to and after feature selection for the test dataset relationship prediction. Results are computed using the annotated concepts (columns 1 and 2), and using the predicted concepts as identified in the concept recognition step (columns 3 and 4).
| Results of relationship identification | Annotated concepts | Predicted concepts | ||
|---|---|---|---|---|
| Before feature selection | After feature selection | Before feature selection | After feature selection | |
| Improves | 0.735 | 0.727 | 0.558 | 0.728 |
| Worsens | 0.598 | 0.600 | 0.398 | 0.541 |
| Causes | 0.655 | 0.664 | 0.401 | 0.632 |
| Given | 0.735 | 0.738 | 0.536 | 0.735 |
| Not Given | 0.628 | 0.427 | 0.494 | 0.458 |
| Reveals | 0.814 | 0.823 | 0.625 | 0.821 |
| Conducted | 0.593 | 0.583 | 0.404 | 0.531 |
| Relates | 0.591 | 0.588 | 0.390 | 0.588 |
| 0.712 | 0.711 | 0.516 | 0.704 | |
The F-measures are computed using the i2b2 evaluation package, and the Average is computed weighting each individual F-measure by the number of all positive examples in that particular relationship category.