Literature DB >> 30961580

Attention-based deep residual learning network for entity relation extraction in Chinese EMRs.

Zhichang Zhang1, Tong Zhou2, Yu Zhang2, Yali Pang2.   

Abstract

BACKGROUND: Electronic medical records (EMRs) contain a variety of valuable medical concepts and relations. The ability to recognize relations between medical concepts described in EMRs enables the automatic processing of clinical texts, resulting in an improved quality of health-related data analysis. Driven by the 2010 i2b2/VA Challenge Evaluation, the relation recognition problem in EMRs has been studied by many researchers to address this important aspect of EMR information extraction.
METHODS: This paper proposes an Attention-Based Deep Residual Network (ResNet) model to recognize medical concept relations in Chinese EMRs.
RESULTS: Our model achieves F1-score of 77.80% on the manually annotated Chinese EMRs corpus and outperforms the state-of-the-art approaches.
CONCLUSION: The residual network-based model can reduce the negative impact of corpus noise to parameter learning, and the combination of character position attention mechanism will enhance the identification features of different type of entities.

Entities:  

Keywords:  Attention mechanism; Chinese electronic medical record; Deep residual learning network; Entity relation extraction

Mesh:

Year:  2019        PMID: 30961580      PMCID: PMC6454667          DOI: 10.1186/s12911-019-0769-0

Source DB:  PubMed          Journal:  BMC Med Inform Decis Mak        ISSN: 1472-6947            Impact factor:   2.796


Background

EMR is used by medical staff to record texts, symbols, charts, graphics, data, and other digital information generated by HIS (hospital information system). With the tremendous growth of the adoption of EMR, various sources of clinical information (including demographics, diagnostic history, medications, laboratory test results, and vital signs) are becoming available, which has established EMR as a treasure trove for large-scale analysis of health data. Unstructured medical text in EMR is one kind of narrative data, including clinical notes, surgical records, discharge records, radiology reports, and pathology reports. For the convenience of narration, we use EMR to represent unstructured EMR text in the following. Identifying semantic relations existing among medical concepts in EMRs is of great importance to health-related various applications. These relations are hold between medical problems, tests, and treatments. Table 1 presents two examples of semantic relation, one of which is between medical concept e1=“cold” and e2=“fever” in sentence S1, and the other is between e1=“Head MRI” and e2=“lacunar infarction” in sentence S2.
Table 1

Examples of the relations between medical entities

SentenceRelation
S1: The patient has a cold, feels afever and headache.Disease causes symptoms (DCS)
S2: Head MRI shows lacunar infarction.Test reveals the disease (TeRD)
Examples of the relations between medical entities On account of the importance of this subject, the 2010 i2b2/VA NLP challenge for clinical Records presented a relation classification task focused on assigning relation types between medical concepts in EMRs. Since then medical concept relation classification has being paid attention by more and more researchers. In the traditional natural language processing (NLP) research, semantic relations between named entities can be used for many applications including knowledge graph construction, sentiment analysis, question answering, etc. [1], relation extraction or classification therefore has always been an important issue [2]. In previous open-domain entity relation extraction studies, researchers applied many different traditional machine learning models include Logistic Regression, SVM and CRF to recognize relations [3-7]. Li et al. used CRF model to reduce the space of possible label sequences and introducing long range features for relation recognition [8]. Mintz et al. put forward a remote monitoring relation classification method which could generate adequate training data by aligning text and knowledge base to solve the problem of lack of enough training data [9]. Socher et al. firstly employed recurrent neural network (RNN) on the task of relation extraction, while utilizing the syntactic structure information of sentences [10]. Miwa et al. proposed a neural network relation extraction architecture based on bidirectional LSTM and tree LSTM to encode entities and sentences simultaneously [11]. Drawing on these studies on open-domain relation extraction, similar task on EMRs was formally defined in the 2010 i2b2/VA Challenge Evaluation [12]. Some researchers proposed various models for relation classification of EMRs. Bruijn et al. used SVM to train multiple classifiers to deal with different relation categories, and improved the effect of classification [13]. Rink et al. use external dictionaries to increase the effect of entity relationship recognition [14]. Fang et al. extracted the relations from relevant articles of Chinese herbal medicine based on manually designed rules and created a relation database [15]. Zhou et al. utilized a bootstrapping framework to extract relations from the medical articles and created a knowledge base [16]. Li et al. raised an electronic health records relation classification model based on CNN-LSTM [17]. Overall, the existing models mainly focus on English EMR texts, and on the other hand it still cannot deliver satisfactory recognition performance. Concerning the increasing availability of digitalized Chinese EMRs, this paper addresses the semantic relation identification problem among medical concepts in Chinese EMRs. We propose an attention mechanism based deep residual network model to classify the medical entity relations in Chinese EMRs. Experimental results performed on a manually labeled Chinese EMR corpus show that our model achieved better performance with F1-score of 77.80% compared with other methods.

Methods

Our model is based on a CNN architecture as shown Fig. 1. The model consists of five parts: vector representation layer, convolution layer, residual networks layer, position attention layer and output layer.
Fig. 1

The architecture of our relation extraction model

The architecture of our relation extraction model

Character embedding

Given a Chinese sentence S=(c1,c2,…,c) which contains two entities e1 and e2. Each character c will be mapped to a low-dimensional dense vector , in which represents the character vector and is the vector of character position in the sentence. The character embedding initialized with vector which is pre-trained by word2vec, and d is the dimension of character vector.

Position embedding

Position embedding is also a low-dimensional vector of character position in the sentence, which can combine the relative positions (see Fig. 2) of the current character to the first entity e1 as well as the second entity e2. Each relative position corresponds to a position embedding , d is the dimension of position embedding.
Fig. 2

An example of the relative distance between an entity and a character. The relative distance of a character to medical entity “(cold)” and “(fever)” are 2 and -2 respectively

An example of the relative distance between an entity and a character. The relative distance of a character to medical entity “(cold)” and “(fever)” are 2 and -2 respectively The vector is concatenation of character vector and two position vectors, where d=d+2d.

Convolution

Convolution is to extract the effective local feature information from characters and their corresponding contexts. The V is a vector which corresponds the j-th character in the sentence S=(V1,V2,…,V), here n is the sentence length. We use filter to extract local features from the sentence S. A feature c is generated from a window of character V by where b is a bias terms and f is a non-linear function. We apply dropout layer in convolution to prevent data from outfitting.

Residual networks

Residual learning connects low-level to high-level representations directly and solves the vanishing gradient problem, we superimposed the identity mapping function on a network. In our model, each residual convolution block (see Fig. 3) has two convolutional layers, each one followed by a ReLU activation, we use shortcut connection between each of the residue convolution block W1,W2∈R are two convolution filters, where h is convolution kernel size. The first convolutional layer is
Fig. 3

The residual convolution block

The residual convolution block and the second is here b1, b2 are bias terms. The residual convolution block output is the vector . This block will be multiply concatenated in our architecture by a shortcut connection.

Position attention

Recently attention mechanism has been widely used in machine learning, and great achievements have been made in various NLP problems. In this paper, we use the position attention to enhance relation extraction ability. Firstly, we carry the max-pooling operation on the residual learning result. Secondly, as shown in Fig. 1, we concatenate the max-pooling results with the position embedding of entity. Finally, we use the attention mechanism to balance the weight to the sentence. where α represents the attention weight. P is a result which concatenates the max-pooling results with the position embedding of entity. Finally, we use the softmax function to normalize and output entity relation probability.

Results

Dataset and evaluation metrics

On the basis of reference to medical semantic relation annotation specification of 2010 i2b2/VA Challenge, we established our own relation annotation specification of Chinese EMRs, in which semantic relations between medical concepts fall into five coarse-grained categories and fifteen fine-grained categories. All of relation category are detailed as follows. Coarse-grained category 1: Treatment -Disease Relation. This category contains five fine-grained categories, including TrID (Treatment improves the disease), TrWD (Treatment worsens the disease), TrCD (Treatment causes the disease), TrAD (Treatment is administered for the disease), and TrNAD (Treatment is not administered because of the disease). Coarse-grained category 2: Treatment -Symptoms Relation. This category also contains five fine-grained categories, including TrIS (Treatment improves the symptoms), TrWS (Treatment worsens the symptoms), TrCS (Treatment causes the symptoms), TrAS (Treatment is administered for the symptoms), and TrNAS (Treatment is not administered because of the symptoms). Coarse-grained category 3: Test-Disease Relation. This category contains two fine-grained categories, including TeRD (Test reveals the disease) and TeCD (Test conducted to investigate the disease). Coarse-grained category 4: Test-Symptoms Relation. This category also contains two fine-grained categories, including TeRS (Test reveals the symptoms) and TeBS (Test based on symptoms). Coarse-grained category 5: Disease-Symptoms Relation. This category contains only one fine-grained category named as DCS (Disease causes symptoms). According to our specification, we manually annotated 3000 de-identified Chinese EMR texts from different clinical departments of a grade-A hospital of second class in Gansu Province, China. 2000 medical texts are selected as training data, 500 medical texts as develop data, and 500 medical texts for test while evaluating our method on this dataset. The relation numbers of every fine-grained category in this dataset are given in Table 2. Precision, Recall and F1-score are used as evaluation metrics.
Table 2

The relation number of every fine-grained category in the corpus

Fine-grained categoryTraingDevelopTest
TrID368260193
TrWD229149102
TrCD423284265
TrAD470620961581
TrNAD1103541
TrIS1351371427
TrWS598152163
TrCS1184157
TrAS209310831154
TrNAS983621
TeRD1770498603
TeCD852327
TeRS1396381804998
TeBS1492214388
DCS546526773251
The relation number of every fine-grained category in the corpus

Models and parameters

We carry out the experiments to compare the performance of our model with others described in the following. : This model was used by Sahu, et al. [18], which encoded the sentence vectors with CNN, and outputted the results after max-pooling and softmax function. : This model was proposed by Li, et al. It mainly consists of bidirectional LSTM and attention mechanism [19]. : This model was proposed by Huang, et al. Compared with our model, this model did not combined attention mechanism [20]. : The basic framework of the method is close to our model. The difference between this one with ours is that this model combine the residual network with Bi-LSTM. : This is the model presented in this paper. Table 3 gives the chosen hyper-parameters for all experiments. We tune the hyper-parameters on the development set by random search. We try to share as many hyper-parameters as possible in experiments.
Table 3

Hyper parameters of the residual neural network

ParameterDescriptionValue
d w Dimension of word embedding100
d p Dimension of position embedding5
k Window size3
m Number of filters128
B Batch size50
λ Learning rate0.01
p The ration of dropout0.3
Hyper parameters of the residual neural network

Experimental results

Table 4 shows the overall classification performance of different models on our evaluation corpus. It can be seen that our method ResNet-PAtt is better than other methods in F1-score while precision, recall and F1-score reaches 79.16 and 77.80% respectively. Of all other methods, the model ResNet-BLSTM achieves the best performance on F1-score, and our model improves 2.97% F1-score compared with it, then our method is more effective. In addition, we can find that overall the residual network based methods are better than other relation extraction methods.
Table 4

Comparison of overall relation classification result of different model

ModelPrecisionRecallF1-score
SVM65.2455.2659.84
CNN-Max [18]55.3450.8452.99
LSTM-Max69.2770.5169.88
BLSTM-Attention [19]74.1266.9570.35
ResNet-BLSTM 78.81 71.2474.83
ResNet-Max [20]65.2467.4566.33
ResNet-PAtt 76.48 79.16 77.80
Comparison of overall relation classification result of different model

Discussion

The reasons our model achieves best performance maybe owe to that the residual network-based model could reduce the negative impact of corpus noise to parameter learning, and the combination of character position attention mechanism could enhance the identification information of different type of entities. Table 5 gives the classification performance of our model on every fine-grained relation category. As can be seen from these data, our model performs best on relation category TeRS and worst on category TrNAS, which shows that it is more difficult to recognize category TrNAS correctly. We also evaluate the training time of different models. Figure 4 shows that the consumed times by these models while epoch is set as 5, 10 and 20 respectively. Overall, our model takes the shortest time to complete parameter training, and the traditional machine learning method SVM takes the longest time to train.
Table 5

Classification performance of our model on every fine-grained relation category.

RelationPrecisionRecallF1-score
TrID46.8442.6144.62
TrWD41.3540.1240.73
TrCD47.7345.3346.50
TrAD72.4268.4870.39
TrNAD45.8846.1846.03
TrIS57.4255.6756.53
TrWS50.2148.2849.23
TrCS38.3642.6440.39
TrAS61.3880.5569.67
TrNAS35.7636.5136.13
TeRD74.8172.5173.64
TeCD41.5539.3540.42
TeRS83.5781.6883.61
TeBS56.7258.3157.50
DCS76.8674.5375.68
Fig. 4

Comparison of the training time for different model

Comparison of the training time for different model Classification performance of our model on every fine-grained relation category. Table 6 is comparison of F1-score for each model on every fine-grained relation category. The model has better classification performance and faster response speed.
Table 6

Comparison of F1-score for each model on every fine-grained relation category

CategorySVMCNN-MaxLSTM-MaxBLSTM-AttentionResNet-BLSTMResNet-MaxResNet-PAtt
TrID20.0629.6836.7840.3842.6735.4244.62
TrWD19.3528.3425.3535.2133.2130.4340.73
TrCD28.5228.0239.4146.3248.5742.646.50
TrAD63.2143.5258.3171.6568.3364.5470.39
TrNAD12.3622.4618.2436.6937.2635.4246.03
TrIS57.2448.5249.3154.3752.4452.3156.53
TrWS36.4149.5137.5346.2148.1842.4349.23
TrCS39.0439.5341.5239.4640.9339.540.39
TrAS60.2658.3362.3466.8372.3661.3669.67
TrNAS13.5414.3914.5628.6730.3124.6736.13
TeRD62.3560.2762.2471.3674.2269.9673.64
TeCD12.3416.5218.3637.2332.8831.4840.42
TeRS82.5371.2675.3480.4481.6378.4583.61
TeBS48.4246.3447.2158.6457.9451.2057.50
DCS64.2865.3165.6774.2473.5570.6975.68
Comparison of F1-score for each model on every fine-grained relation category

Conclusions

In this paper, we propose a deep residual network model based on the attention mechanism to classify the relation of entity pairs in Chinese EMRs. The method reduced the influence of data noise on the model training, and enhance entity discrimination feature with position attention mechanism so that the entity information can be combined effectively in the relation extraction. Experimental results show that the model reached 77.80% F1-score value, and significantly improved the classification performance of the few instance categories. At present, most relation classifications are based on entity recognition tasks and need to specify the entity in the sentence. In the future, we will study the joint extraction of entity and entity relation to further improve the efficiency of entity and entity relation recognition simultaneously.
  5 in total

1.  2010 i2b2/VA challenge on concepts, assertions, and relations in clinical text.

Authors:  Özlem Uzuner; Brett R South; Shuying Shen; Scott L DuVall
Journal:  J Am Med Inform Assoc       Date:  2011-06-16       Impact factor: 4.497

2.  Automatic extraction of relations between medical concepts in clinical texts.

Authors:  Bryan Rink; Sanda Harabagiu; Kirk Roberts
Journal:  J Am Med Inform Assoc       Date:  2011 Sep-Oct       Impact factor: 4.497

3.  Integrative mining of traditional Chinese medicine literature and MEDLINE for functional gene networks.

Authors:  Xuezhong Zhou; Baoyan Liu; Zhaohui Wu; Yi Feng
Journal:  Artif Intell Med       Date:  2007-09-05       Impact factor: 5.326

4.  Machine-learned solutions for three stages of clinical information extraction: the state of the art at i2b2 2010.

Authors:  Berry de Bruijn; Colin Cherry; Svetlana Kiritchenko; Joel Martin; Xiaodan Zhu
Journal:  J Am Med Inform Assoc       Date:  2011-05-12       Impact factor: 4.497

5.  TCMGeneDIT: a database for associated traditional Chinese medicine, gene and disease information using text mining.

Authors:  Yu-Ching Fang; Hsuan-Cheng Huang; Hsin-Hsi Chen; Hsueh-Fen Juan
Journal:  BMC Complement Altern Med       Date:  2008-10-14       Impact factor: 3.659

  5 in total
  7 in total

Review 1.  Deep learning in clinical natural language processing: a methodical review.

Authors:  Stephen Wu; Kirk Roberts; Surabhi Datta; Jingcheng Du; Zongcheng Ji; Yuqi Si; Sarvesh Soni; Qiong Wang; Qiang Wei; Yang Xiang; Bo Zhao; Hua Xu
Journal:  J Am Med Inform Assoc       Date:  2020-03-01       Impact factor: 4.497

2.  A hybrid method based on semi-supervised learning for relation extraction in Chinese EMRs.

Authors:  Chunming Yang; Dan Xiao; Yuanyuan Luo; Bo Li; Xujian Zhao; Hui Zhang
Journal:  BMC Med Inform Decis Mak       Date:  2022-06-27       Impact factor: 3.298

3.  Construction and Application of Text Entity Relation Joint Extraction Model Based on Multi-Head Attention Neural Network.

Authors:  Yafei Xue; Jing Zhu; Jing Lyu
Journal:  Comput Intell Neurosci       Date:  2022-05-24

4.  Text Semantic Classification of Long Discourses Based on Neural Networks with Improved Focal Loss.

Authors:  Dan Jiang; Jin He
Journal:  Comput Intell Neurosci       Date:  2021-01-07

5.  Multi-task learning for Chinese clinical named entity recognition with external knowledge.

Authors:  Ming Cheng; Shufeng Xiong; Fei Li; Pan Liang; Jianbo Gao
Journal:  BMC Med Inform Decis Mak       Date:  2021-12-31       Impact factor: 2.796

6.  Named Entity Recognition of Medical Text Based on the Deep Neural Network.

Authors:  Tianjiao Yang; Ying He; Ning Yang
Journal:  J Healthc Eng       Date:  2022-03-07       Impact factor: 2.682

7.  Noninvasive detection and interpretation of gastrointestinal diseases by collaborative serum metabolite and magnetically controlled capsule endoscopy.

Authors:  Xiang-Tian Yu; Ming Chen; Jingyi Guo; Jing Zhang; Tao Zeng
Journal:  Comput Struct Biotechnol J       Date:  2022-10-06       Impact factor: 6.155

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.