Literature DB >> 32432096

Protein Interaction Network Reconstruction Through Ensemble Deep Learning With Attention Mechanism.

Feifei Li1, Fei Zhu1,2, Xinghong Ling1, Quan Liu1.   

Abstract

Protein interactions play an essential role in studying living systems and life phenomena. A considerable amount of literature has been published on analyzing and predicting protein interactions, such as support vector machine method, homology-based method and similarity-based method, each has its pros and cons. Most existing methods for predicting protein interactions require prior domain knowledge, making it difficult to effectively extract protein features. Single method is dissatisfactory in predicting protein interactions, declaring the need for a comprehensive method that combines the advantages of various methods. On this basis, a deep ensemble learning method called EnAmDNN (Ensemble Deep Neural Networks with Attention Mechanism) is proposed to predict protein interactions which is an appropriate candidate for comprehensive learning, combining multiple models, and considering the advantages of various methods. Particularly, it encode protein sequences by the local descriptor, auto covariance, conjoint triad, pseudo amino acid composition and combine the vector representation of each protein in the protein interaction network. Then it takes advantage of the multi-layer convolutional neural networks to automatically extract protein features and construct an attention mechanism to analyze deep-seated relationships between proteins. We set up four different structures of deep learning models. In the ensemble learning model, second layer data sets are generated with five-fold cross validation from basic learners, then predict the protein interaction network by combining 16 models. Results on five independent PPI data sets demonstrate that EnAmDNN achieves superior prediction performance than other comparing methods.
Copyright © 2020 Li, Zhu, Ling and Liu.

Entities:  

Keywords:  attention mechanism; deep learning; ensemble learning; multi-layer convolutional neural network; protein-protein interaction; protein-protein interaction network

Year:  2020        PMID: 32432096      PMCID: PMC7215070          DOI: 10.3389/fbioe.2020.00390

Source DB:  PubMed          Journal:  Front Bioeng Biotechnol        ISSN: 2296-4185


Introduction

Protein interactions and interaction networks take part in vital activities of each living cell, including signal transduction, immune response, metabolism of energy substance, cell cycle control, etc. (Keskin et al., 2016). The exact identification of protein interactions is therefore important not only to understanding the functions of proteins but also to structure-based drug design and treatment of diseases (Li et al., 2009). Majority of existing methods for predicting PPI are based on Gene Ontology and annotations, phylogenetic profile, gene fusion, the interacting proteins co-evolution pattern and the similarity of proteins in sequence, structure, domain and subcellular localization (Boxem et al., 2008; Zhang et al., 2012; Planas-Iglesias et al., 2013; Sun et al., 2017). However, as their accuracy and reliability depend heavily on collected prior knowledge, they are hardly applied widely. Several methods based on amino acid sequence computation have been explored to predict PPI, such as support vector machine with traditional auto-correlation, k-nearest neighbor (kNN) with local description(LD) (Yang et al., 2010), support vector machine (SVM) with conventional auto covariance(AC) (Guo et al., 2008) or local description(LD) (Zhou et al., 2011), deep neural network with amphiphilic Pseudo amino acid composition (PseAAC) descriptor (Du et al., 2017b) and so on. The above methods provide different techniques of protein sequences such as AC, LD, MCD, PseAAC, with each technique extracting different feature information of protein interactions (Zhang et al., 2019a). AC and CT considered the physical properties of amino acids and their dipole and side-chain volumes respectively. Then LD uses triples to describe composition, transition and distribution of sequence, while PseAAC further studies order information of sequences. We propose to combine different descriptors to achieve PPI prediction to obtain more information from protein interactions. Ensemble learning is a machine learning method, which uses a series of learners and uses some rules to integrate the learning results so as to obtain better performance than a single learner. And ensemble learning has broad application prospects in many fields such as protein phosphorylation site prediction, genome function prediction and cancer prediction in bioinformatics (Gomes et al., 2017; Krawczyk et al., 2017). The previous works also demonstrate the effectiveness of classifier ensemble and provide some guidelines to generate an ensemble classification model (Martin et al., 2005; Han and Huang, 2006; Huang and Zheng, 2006; Huang and Du, 2008). Wang used a boosting technique to generate multiple classifiers iteratively to solve the problem of imbalance between positive and negative data when predicting the phosphorylation sites (Wang et al., 2017). Wang took a random forest and voting method as a basic classifier integration strategy separately to predict PPI sites (Wang et al., 2019). You et al. (2019) chose the basic classifiers with optimal performance, leaving the classifiers with small differences and using the max-wins voting (MWV) strategy to predict DNA binding proteins. Zhang et al. (2019a) trained 27 models by combining AC, MCD, LD with 9 DNN models of different configurations, and integrated these models through Double-layers BP Neural Network. Furthermore, when exploring protein interactions and interaction networks, it is nonnegligible to quantify the interaction/non-interaction relationship between two proteins. One solution is to directly concatenate the features of the two proteins to form a feature vector (Zhang et al., 2019b), which lacks the information characteristics of the interaction/non-interaction between two proteins; another solution is to extract two features with two different networks and combine the features to form a new feature vector as the input of the model (Du et al., 2017b; Hashemifar et al., 2018), which is incapable of learning inherent relation of the proteins. Recently in natural language processing domain, researches have shown that attention mechanisms can effectively emphasize the relatively important parts of the input sentences and help boost the performance of relation extraction (Chen et al., 2017; Du et al., 2017a). In bioinformatics, attention mechanism is also adopted in chemical-protein interaction (CPI) (Zhang et al., 2019b), kinase-specific phosphorylation site prediction (Wang et al., 2017) and so on. In Xuan et al. (2019) model, exploiting the attention mechanism module to learn features or extract the relationship between IncRNA and disease provides more information. Wang et al. (2017) designed a two-dimensional independent attention mechanism for predicting phosphorylation sites, which enabled the model, called MusiteDeep, to automatically search important positions of the output sequences to estimate the contribution of each element in the sequences and feature dimensions. However, the above researches concern only single attention mechanism in the deep neural network model, which can be replaced by the multi-head attention mechanism that can exert attention multiple times and divide attention information into multiple heads. Liu et al. (2018) integrated attention pooling into the gated recurrent unit (GRU) model to extract CPIs. Verga et al. (2018) combined the Multi-head attention with convolution neural networks to construct a transformer model to extract the document-level biomedical relations. Thus, a multi-head attention mechanism will make it easier to capture the relevant important information for deep neural networks in PPI extraction. Motivated by attention mechanisms and ensemble learning, we propose an algorithm called EnAmDNN, which at first extracted the biophysical-chemical information of protein sequences through AC, CT, LD, and PseAAC and association with the interactive description of each protein in protein interaction network; then it automatically extracted the protein features by multi-layer convolutional neural network, adopted attention mechanism to analyze deep-seated relationship of proteins and then forms the feature vectors. In EnAmDNN, 16 kinds of DNN models are trained through 4 characteristic bases which are the inputs of 4 DNNs with different layers and different neurons. In the integration module, the outputs of 16 DNNs are taken as the inputs of deep neural networks finally, and the five-fold cross validation is adopted to comprehensively predict protein interactions and interaction networks. Our contributions can be summarized as follows: (1) the new network structure can automatically extract highly abstract representations and detect the sequence specificity of proteins; (2) the attention mechanism is adopted to analyze internal links between the two proteins and the network description of each protein, instead of directly concatenating the two proteins, to improve the prediction accuracy;(3) ensemble learning considers the advantages of different descriptors and different DNNs to achieve comprehensive learning.

Preliminaries

Deep Neural Network

It turns out that deep neural network (DNN) plays an important role in bioinformatics (Alipanahi et al., 2015; Zhou and Troyanskaya, 2015; Liu et al., 2016), i.e., predicting inner-organization and trans-organization RNA splicing patterns (Leung et al., 2014). DeepMind applied DNN to the detection of sequence specificity of the DNA-RNA binding protein, which is superior to other methods (Alipanahi et al., 2015); DeepSEA applied DNN to learn the code of regulatory sequences from chromatin map sequences in order to discern priorities of other functional varieties (Zhou and Troyanskaya, 2015); other examples include genome informatics extraction, detection of protein structure and medicine discovery. In short, compared with other sequence-based methods, DNN has the following advantages: (1) it can automatically learn certain protein sequences; (2) it can reduce the influence of noise on the raw data and extract the hidden high dimension representation (Bengio et al., 2013). However, the performance of DNN is closely related to the network configuration and may vary greatly for different configurations.

Protein Representation Technique

Different representation techniques of protein features may have a strong impact on the performance of PPI prediction, making it a challenge to effectively express the protein features and describe the connections of two proteins. We choose four representative protein techniques instead of one to avoid the limitation brought by a single technique.

Auto Covariance Technique

Two proteins interact with each other through electrostatic, hydrophobic, steric and hydrogen bond, which can be reflected by the seven physicochemical properties of amino acids, including hydrophobicity (H1), hydrophilicity (H2), volumes of side chains of amino acids (VSC), polarity (P1), polarizability (P2), solvent-accessible surface area and net charge index of side chains. The above properties are exploited by the auto-covariance method to transform amino acid sequence into uniform matrices which reveal the special connection of two residues under a certain distance and are widely applied in protein- encoding. For example, a protein sequence of length L is calculated as follows (Guo et al., 2008): Xij represents the j-th physical property of the i-th amino acid in the protein sequence; lag represents the distance between residues; then proteins of various lengths are encoded as vectors of equal length lg * p, where lg is the maximum lag (lag = 1, 2, …, lg), p is the number of physical properties. In this study, p was 7, reflecting the characteristics of the seven amino acids. As with Guo, we set the log to 30 (Guo et al., 2008). Therefore, each protein sequence is represented as a 210-dimensional vector.

Conjoint Triad Technique

Shen et al. (2007) introduced a conjoint triad technique to represent sequence information of each protein, in which any three contiguous amino acids are regarded as a unit and the characteristics of one amino acid and its vicinal amino acids are fully considered. First, the conjoint triad divides 20 standard amino acids into 7 groups according to their dipole and side-chain volumes, then the triads can be distinguished according to the type of amino acid. According to Shen's settings, there are 343 (7 × 7 × 7) triad types (Shen et al., 2007), as shown in Figure 1.
Figure 1

For a grouped sequence “2762247,” the numerical code string of consecutive amino acids are “276,” “762,” “622,” “224,” “247,” and “*27,” “47*” according to Shen, and “*” is considered to be the first or second amino acid of an amino acid in a continuous amino acid. So its triad types are F276, F762, F622, F224, F427.

For a grouped sequence “2762247,” the numerical code string of consecutive amino acids are “276,” “762,” “622,” “224,” “247,” and “*27,” “47*” according to Shen, and “*” is considered to be the first or second amino acid of an amino acid in a continuous amino acid. So its triad types are F276, F762, F622, F224, F427. Finally, the PPI information of protein sequences are projected into the homogeneous vector space according to the frequency of each triad type, where each protein is represented by a 343-dimensional vector.

Local Descriptor Technique

The Local descriptor technique (Zhou et al., 2011) also divided 20 standard amino acids into 7 groups as shown in Table 1 and divided the entire protein sequence into 10 regions as shown in Figure 2. For each sub-sequence, three descriptors, composition (C), transition (T) and distribution (D), are applied to describe its trait where C represents the proportion of each amino acid group; T represents the frequency with which amino acids in one group are followed by amino acids of another group; D measures the proportion of chain length where the top 25, 50, 75, and 100% of the amino acids of a particular group are located. For the local descriptor method, each region produces 63 values, where C represents 7, T represents 7, and D represents 35, and then each protein is encoded as a 630-dimensional vector.
Table 1

Division of amino acid into seven groups based on the dipoles and volumes of the side chains.

Group 1Group 2Group 3Group 4Group 5Group 6Group 7
A, G, VCF, I, L, PM, S, T, YH, N, Q, WK, RD, E
Figure 2

Ten regions (A–J) of the entire protein sequence. The regions (A–D) are generated by dividing the whole sequence into four equal regions, and regions (E,F) are generated by dividing the whole sequence into two equal regions. The region (G) stand for the central 50% of the entire sequence. And the regions (H–J) stand for the first 75%, the final 75% and the central 75% of the entire sequence respectively.

Division of amino acid into seven groups based on the dipoles and volumes of the side chains. Ten regions (A–J) of the entire protein sequence. The regions (A–D) are generated by dividing the whole sequence into four equal regions, and regions (E,F) are generated by dividing the whole sequence into two equal regions. The region (G) stand for the central 50% of the entire sequence. And the regions (H–J) stand for the first 75%, the final 75% and the central 75% of the entire sequence respectively. For example, according to Table 1, the sequence “ACLACLCCLAALLCCCLALALAAALL” is converted into the amino acid group “1231232231133232131131311133133” so that the sub-sequence contains 9 “1”, 7 “2,” and 10 “3.”For feature C, 9/(9 + 7 + 10) = 0.3461, 7/(9 + 7 + 10) = 0.2693, 10/(9 + 7 + 10) = 0.3846; for feature T, there are 2 cases that “1” is converted to “2” or “2” is converted to “1,” then 2/25 = 0.08; similarly, transitions between “3” and “1” as well as “2” and “3” are 3/25 = 0.12, 6/25 = 0.24, respectively; for feature D, there are nine “1”s, then the D descriptor for 1 is 1/26 = 0.0384, [0.25*9 + 0.5]/26 = 0.0769, [0.5*9 + 0.5]/26=0.1923, [0.75*9 + 0.5]/26 = 0.2692, [1*9 + 0.5]/26 = 0.3462.

Pseudo Amino Acid Composition (PseAAC) Technique

Tian et al. (2019) used a sequence encoding technique based on pseudo amino, that is, PseAAC. Given a protein sequence P with L amino acid residues: where S represents the ith residue of the protein P, 1 ≤ i ≤ L. According to the PseAAC technique, the protein P can be formulated as where the 20 + λ components are given by In Equation (3), f(k = 1, 2, …, 20) are the normalized occurrence frequencies of 20 amino acids in protein P; ω is the weighting factor set to 0.05 in general work; and θ(j = 1, 2, …, λ) denotes the order relationship between two residues that are j residues apart, which is shown as follows: where Jdenotes the order relationship function between amino acid A and A, H(A) denotes the pth property of A. H1(A),H2(A) and H3(A) are the hydrophobicity value, hydrophilicity value and side-chain mass for the amino acid, respectively. This coding method contains more sequence characteristics because it considers not only the physicochemical properties of the protein but also the order information of sequences.

Materials and Methods

Data Sets

We collect the dataset information of Parkinson's disease (PD), Alzheimer's disease (AD), cancer, cardiac and diabetes, whose interactive information is from IntAct database (Kerrien et al., 2007) and sequence information from Uniprot (Bairoch et al., 2004). We are concerned about positive-negative selection in our work. For the positive set, proteins and protein pairs that contain less than 50 amino acids and 40% of sequence identity are removed to eliminate the variance caused by minor bias proteins to the model. The negative set was obtained by pairing proteins whose subcellular localization is different (Guo et al., 2008) or GO Cellular Component (CC) and Biological Process (BP) ontology with experimental evidence codes (Muley and Ranjan, 2012). The subcellular location information on the proteins is extracted from Uniprot. According to this information, a protein can be divided into several types of localization cytoplasm, nucleus, mitochondrion, endoplasmic reticulum, Golgi apparatus, peroxisome and vacuole. The way to construct negative set must meet the following requirements: (1) the non-interacting pairs cannot appear in the positive data set; (2) the contribution of proteins in the negative set should be as harmonious as possible. In our work, the ratio between positive and negative set is 1:1, where the negative sets are randomly chosen from non-interactive pairs. Finally, we have five independent PPI datasets: Parkinson's disease (PD) (4,127 interacting pairs and 4,127 non-interacting pairs), Alzheimer disease (AD) (4,096 interacting pairs and 4,096 non-interacting pairs), Cancer (6,352 interacting pairs and 6,352 non-interacting pairs), Cardiac (2,663 interacting pairs and 2,663 non-interacting pairs) and Diabetes (163 interacting pairs and 163 non-interacting pairs).

Evaluation Criteria

The following metrics are taken into account to perform evaluation: Overall Prediction Accuracy, Recall, Precision, F1 score values, and Area under the ROC Curve (AUC) (Zhang et al., 2019a). The first four metrics are defined as follows: where TP (true positive) is the number of true interacting pairs found in the positive data set, TN (true negative) is the number of true non-interacting pairs with correct prediction, FP (false positive) is the number of the predicted interacting pairs not found in the positive data set, and FN (false negative) is the number of the true interacting pairs with false prediction.

Ensemble Deep Neural Networks

This section describes EnAmDNN model that predicts PPI based on protein sequences, which consists of the input module, the convolution module, the attention mechanism module, the DNN module and the integration module. Each protein sequence is encoded, by the input module, through the protein representation techniques, as a vector, whose feature is extracted by the convolution module. Then, the internal link in the protein pair is detected through the attention module, and then the analyzed protein pair is provided to 16 dependent learners. After the training is completed, these learners will be integrated through a two-layer hidden layer neural network. The working process of an EnAmDNN is shown as Figure 3.
Figure 3

Flowchart of EnAmDNN for predicting protein-protein interactions. First, the interaction pairs and non-interaction pairs of related proteins are obtained from IntAct, and all protein sequence data of UniProt are obtained; the appropriate proportion of interaction pairs and non-interaction pairs are selected, and each group of protein pairs (including interaction pairs and non-interaction pairs) is vectorization by AC, CT, LD, and PseAAC techniques; put vector protein into convolution neural network for feature extraction of each protein; extracted features are transferred to attention mechanism module for deep analysis of interaction between each group of protein pairs; then the analyzed features are input into deep neural network of different models for training; finally, the final prediction results are obtained by integrating the prediction results of different models.

Flowchart of EnAmDNN for predicting protein-protein interactions. First, the interaction pairs and non-interaction pairs of related proteins are obtained from IntAct, and all protein sequence data of UniProt are obtained; the appropriate proportion of interaction pairs and non-interaction pairs are selected, and each group of protein pairs (including interaction pairs and non-interaction pairs) is vectorization by AC, CT, LD, and PseAAC techniques; put vector protein into convolution neural network for feature extraction of each protein; extracted features are transferred to attention mechanism module for deep analysis of interaction between each group of protein pairs; then the analyzed features are input into deep neural network of different models for training; finally, the final prediction results are obtained by integrating the prediction results of different models.

Deep Convolutional Module

The convolution module is a batch of normalized layers, a stack of convolutional layers and activation layers, which can automatically extract features of vectorized protein sequences. In our model, the output of the convolution module is calculated by an expression that starts with a convolution layer and ends with a convolution layer: Repeat In n times and enter the result In into Equation 11. Where Out is the output vector, In is the input vector and β, γ, δ, λ are the parameters of batch normalization and convolution layers. The convolution layer searches the sequences according to their input orders and outputs the corresponding features; the batch normalization layer takes in the feature vectors and normalizes their mean values and the variances; ReLU layer takes in the normalized vectors and introduces non-linearity to achieve complex function approximation. Then the above processes repeat n times to obtain the feature vector.

Attention Module

Convolution layer can automatically learn potential features from input sequences, but only a small part of these potential features are very important in PPI. In our model, we use the multi-attention mechanism to adjust the weight of the input sequences to further emphasize the relatively crucial features. Applying the attention multiple times may learn more important features than single attention and allowing the model to learn relevant information in different representation subspaces (Vaswani et al., 2017). It can be understood that attention selectively selects a small amount of important information that is beneficial to PPI from a large amount of information and focuses on important information, ignoring most of the insignificant information. We choose the mechanism of multiple attention rather than directly connecting the two protein eigenvectors to increase the exploration of protein pairs and further use the indirect relationship between residues to obtain more accurate information. The Multi-head attention calculates the output based on the query and a set of key-value pairs, where Q, K, V denote query, key, and value respectively. The specific structure is shown in Figure 4:
Figure 4

(A) Multi-Head Attention consists of several attention layers. (Vaswani et al., 2017) First, query, key and value go through a linear transformation, and then enter them into scaled dot-Product attention to generate many heads; then concatenate these heads to keep relevant information in different representation subspaces. (B) Scaled dot-product attention (Vaswani et al., 2017). Obtain weights by similarity calculation between query and each key, and the weights are normalized by softmax function; then attention is obtained by the weight and the corresponding value.

(A) Multi-Head Attention consists of several attention layers. (Vaswani et al., 2017) First, query, key and value go through a linear transformation, and then enter them into scaled dot-Product attention to generate many heads; then concatenate these heads to keep relevant information in different representation subspaces. (B) Scaled dot-product attention (Vaswani et al., 2017). Obtain weights by similarity calculation between query and each key, and the weights are normalized by softmax function; then attention is obtained by the weight and the corresponding value. Query, key and value go through a linear transformation first, and then enter them into scaled dot-Product attention. At this time, the attention calculation formula is as follows: Where is scaling factor. The core of the Multi-head attention is employing the above attention multiple times, and one time attention means one head. Suppose the Multi-head attention needs to be done h times to generate h heads, the Atthead can be calculated as follows: where are parameter matrices. Finally, these heads are concatenated and once again linearly transformed by In order to keep the invariance of features, we introduce average pooling and maximum pooling to reduce the errors caused by model parameters and retain information of global and local features. where AvgPool is the function of average pooling and MaxPool is the function of maximum pooling. For a protein pair (P1, P2), it is expressed as S1, S2 respectively after convolution layer. We use the merge layer to fuse the protein pairs that are redistributed by the attention mechanism. The calculation formula of the merge layer is as follows: where A is weight. The basic model of our work is mainly constructed by 3.3.1 the deep convolution module, 3.3.2 the attention mechanism module and the deep neural network. The specific basic learning algorithm is shown in algorithm 1. Basic learning algorithm based on multi-head attention with pooling 1: for i = 1 to n: 2: S1, S2 ← x(s, s) represent encoded protein p1 and protein p2 3: for iter = 1 to t: 4: S1, = RELU (Batchβ, δ(Convλ(S1, ))) 5: S2, = RELU (Batchβ, δ(Convλ(S2, ))) 6: end for 7: S1 ← Convγ(S1, ) 8: S2 ← Convγ(S2, ) 9: calculate the importance value of single attention by 10: connect multiple attentions by , 11: Calculate new feature importance value of multi-head attention with pooling by 12: merge the sequence feature pairs as the input of the network by 13: output = softmax(W · merge + b) 14: end for 15: Return W, b

Ensemble Strategy

Ensembles of independent deep neural networks can improve the performance of a single network (Bairoch et al., 2004). In the Otto group product classification challenge, the first one won the championship by stacking 30 models. The model achieved remarkable results, and we also adopted the stacking method of ensemble learning in our work. Secondly, to predict the effect better, the trainers of each primary model keep stability and diversity as much as possible. We modify the internal structure of the algorithm and learn from different feature representations, which are two strategies to maintain diversity and achieved improvement, so we also take the same measures (Zhang et al., 2019a). In practice, we choose four feature representations to quantify the characteristics of each protein and set different parameters of DNNS according to the characteristics of each representation. Then we use the stacking method to combines with five-fold cross validation, the primary learners are trained from the initial data set, and a new data set is generated by the primary learners for training the secondary learner. It means the output of each primary trainer is input as an example to the secondary trainer for fusion output and PPI prediction. Here, the secondary trainer is composed of deep neural networks. Its structure is shown in Figure 5 and ensemble strategy is described in algorithm 2.
Figure 5

Ensemble strategy composed of deep neural networks. The first layer results (Pn, Tn) (0 < n < t+1) are predicted by T primary learners, where Pn and Tn stand for training data and the prediction result; then use recombined (Pn, Tn) as training data features of the second-level classifier and put it into deep neural networks to predict protein interaction networks.

Ensemble strategy composed of deep neural networks. The first layer results (Pn, Tn) (0 < n < t+1) are predicted by T primary learners, where Pn and Tn stand for training data and the prediction result; then use recombined (Pn, Tn) as training data features of the second-level classifier and put it into deep neural networks to predict protein interaction networks.

Result

Comparing the Prediction Performance With Other Methods

All the experiment were carried out on a computer with CentOs, Cuda version 10.1.243, CuDnn version 7.0 and software environment python3.7+keras2.0+torch1.3. EAM algorithm 1: initialize weight matrix W and Bias b, iter as 1 2: extract protein sequence from uniprot 3: vectorize protein sequence by AC, LD, CT, PseAAC 4: divide training data and testing data TeD 5: for m = 1 to M do 6: for k = 1 to K do 7: train with 8: predict D and require predicted value P = h(D) 9: predict T and and require predicted value T = h(T) 10: end for 11: splicing predicted value of training data through mth basic model by 12: splicing predicted value of testing data through mth basic model by T = (T1 + T2 + … + T)/K 13: end for 14: construct new training data by 15: and construct new testing data by newT = (T1, T2, …, T) 16: Repeat: 17: iter + 1 18: for i = 1 to n: 19: set 20: for l=2 to L: a = σ(Wa + b) 21: calculate probability p divided into class k 22: calculate 22: for l=2 to L: 23: end for 24: for l=2 to L: 25: 26: W ← W − ∇W, b ← b − ∇b 27: end for 27: Until ∇W, ∇b < ε 28: Return W, b In order to evaluate the performance of EnAmDNN, we compared it with the approaches proposed by Guo et al. (2008), Zhou et al. (2011), Du et al. (2017b), Yang et al. (2010), Zhang et al. (2019a) emphasized the highest score of each evaluation criteria in bold and present the results in Tables 2–6, which separately utilize AC, LD, CT, APAAC, PseAAC to encode amino acid sequence, and predicted PPI with SVM, k-nearest neighbor (kNN) or DNNs, all of which share the same training sets and the same testing sets. It can be seen from Table 2 that EnAmDNN generally outperforms these predictors, where EnAmDNN achieved optimal prediction performance in all the datasets, especially in AD, with an accuracy of 94.67%, and a recall rate of 93.29%. The accuracy is 95.41%, F1 is 94.33%, and AUC is 94.63%. This is because, in EnAmDNN, feature representations in protein sequences are coordinated, and new features are obtained through different classifiers. Compared with the recent EnsDNN model, in five independent data sets, the AUC index DnAmDNN has increased by 0.89, 0.41, 0.61, 0.6, 3.90%, and the accuracy of PPI prediction are relatively high. The EnAmDNN model takes advantage of the multi- head attention mechanism, that is, extracts the internal links of the PPI, thereby improving the performance of the model.
Table 2

AD with various methods.

MethodsAccuracyRecallPrecisionF1AUC
EnsAmDNN0.94670.93290.95410.94330.9463
SVM_AC0.78860.63690.89560.74350.7831
kNN_LD0.75490.75070.76020.75420.7556
SVM_LD0.8040.67350.91740.77540.805
NDDs_APAAC0.8980.90930.89070.89930.8977
EnsDNN0.93720.92550.95310.93880.938

The highest score of each evaluation criteria is emphasized in bold.

Table 6

Diabetes with various methods.

MethodsAccuracyRecallPrecisionF1AUC
EnsAmDNN0.83330.8710.79410.83080.8355
SVM_AC0.78910.66670.79410.71280.7819
kNN_LD0.75680.85710.750.80.7411
SVM_LD0.78130.82690.73610.77420.7885
NDDs_APAAC0.8030.91180.75610.82670.7996
EnsDNN0.80300.75760.83330.79370.8030

The highest score of each evaluation criteria is emphasized in bold.

AD with various methods. The highest score of each evaluation criteria is emphasized in bold. PD with various methods. The highest score of each evaluation criteria is emphasized in bold. Cancer with various methods. The highest score of each evaluation criteria is emphasized in bold. Cancer with various methods. The highest score of each evaluation criteria is emphasized in bold. Diabetes with various methods. The highest score of each evaluation criteria is emphasized in bold. To further demonstrate the effect of ensemble strategy, five-fold cross-validation is employed to improve the reliability of the results. Figure 6 shows the performance of each basic learner, where it can be observed, taking AD dataset as an example, that each basic learner, associated with five-fold cross-validation method, shows fairish prediction performance, which is reflected on all the evaluation criteria Accuracy, Recall, Precision, F1, AUC. The result indicates that our model extracts and trains the features produced by basic learners and that the shortcoming of each basic learner is overcome to a certain degree. It's also confirmed with PD, Cardiac and Cancer in Figures 7–9.
Figure 6

Comparison of evaluation indexes of each basic model and ensemble model with AD data set.

Figure 7

Comparison of evaluation indexes of each basic model and ensemble model with Cancer data set.

Figure 9

Comparison of evaluation indexes of each basic model and ensemble model with PD data set.

Comparison of evaluation indexes of each basic model and ensemble model with AD data set. Comparison of evaluation indexes of each basic model and ensemble model with Cancer data set. Comparison of evaluation indexes of each basic model and ensemble model with Cardiac data set. Comparison of evaluation indexes of each basic model and ensemble model with PD data set.

Performance of PPI Prediction

To further study the effectiveness of the ensemble strategy, we designed two different network architectures: (a) concatenating four feature representations (AC, LD, CT, PseAAC) as the input to the first layer classifiers (namely EnAm-Con) and (b) separately taking one feature representation as the input to the first layer classifiers (namely EnAm-Sep including EnsAC, EnsLD, EnsCT, EnsPseAAC). EnAm-Con first concatenates four feature representations and then integrated 12 trained DNNs in the same way as EnAmDNN. For EnAm-Sep, we separately trained 12-model DNNs based AC, LD, CT, and PseAAC, and integrated these DNNs in the same way as EnAmDNN. The performance of EnAm-Con and EnAm-Sep which emphasized in bold are also listed in Table 7 where it can be observed that the LD method performs better than AC and PseAAC method. The LD method of AUC value obtained from the first four data sets are 6.2, 4.89, 3.57, 7.17, 0.8 and are 16.37, 9.05, 14.19, 17.89, 9.46% higher than AC and PseAAC methods separately, which is because LD can encode more interaction information. It can be seen from Table 6 that EnAm-Con performs better than EnAm-Sep, proving that concatenating different feature representations as new feature vector can improve the accuracy of the ensemble strategy. It can be concluded that these four representations are complementary to each other and our ensemble strategy is effective and feasible.
Table 7

Comparison of EnAm-Con and EnAm-Sep.

Data setsEnsACEnsLDEnsCTEnsPseAACEnsConEnAMDNN
AD0.85020.90290.89490.77590.93380.9456
PD0.83780.87880.88200.80590.92420.8951
Cancer0.76380.79110.78890.69280.84680.8508
Cardiac0.82010.87890.88880.74550.90200.9088
Diabetes0.75160.75760.79170.69210.84590.8787

The highest score of each evaluation criteria is emphasized in bold.

Comparison of EnAm-Con and EnAm-Sep. The highest score of each evaluation criteria is emphasized in bold. The number of basic learners greatly influences the overall prediction performance, where the efficiency of the model continues to grow as the number of learners increases, to a point that the performance tends to be stable. To evaluate the influence of the EnAmDNN, we assign different numbers of DNNs to protein represent technique, such as 1, 3, 5, 7, 9. The result is presented in Figure 10, where it can be observed that the AUC of the EnAmDNN tends to be stable when the number reaches 16. The efficiency of the EnAmDNN may also be affected by the performance of each basic learner, for which we prepare 16 basic learners, iterate them for 600 times, and combine them through deep neural network. The result is shown as follows. It can be seen that the prediction performance improves as the iteration continues and the model tends to remain stable at the point of 200.
Figure 10

(A) The influence of the number of basic learners on the EnAmDNN. It can be observed that the AUC of the EnAmDNN tends to be stable when the number reaches 16. (B) The influence of the number of training iterations of the basic learners on the EnAmDNN. It can be seen that the model tends to remain stable at the point of 200.

(A) The influence of the number of basic learners on the EnAmDNN. It can be observed that the AUC of the EnAmDNN tends to be stable when the number reaches 16. (B) The influence of the number of training iterations of the basic learners on the EnAmDNN. It can be seen that the model tends to remain stable at the point of 200. Table 8 reports the process running time of EnAmDNN based on fold cross-validation with 16 basic learners and iterate each basic learner for 600 times.
Table 8

Running time of EnAmDNN based on fold cross-validation.

Date setsADPDCancerCardiacDiabetes
Time (s)1,273,589.80835,292.292,618,297.111,145,029.32101,784.60
Running time of EnAmDNN based on fold cross-validation. Meanwhile, to further investigate the contribution of using an ensemble predictor with fold cross-validation, we integrated the simplified EnAmDNN, which don't use fold cross-validation. To reduce the impact of data dependency in the experiment, we constructed data sets on Cardiac based on Muley and Ranjan (2012) to observe the performance of proposed model. From Table 9, we can see that EnAmDNN achieves competent prediction performance with an average accuracy of 85.66%, precision of 89.47%, F1 of 85.16%, and AUC of 85.76. It has better performance than simplified EnAmDNN across evaluation metrics. The prediction results show that EnAmDNN with fold cross-validation is capable of predicting PPIs.
Table 9

Comparison of EnAmDNN and simplified EnAmDNN.

ModelsAccuracyRecallPrecisionF1AUC
EnAmDNN0.87490.81380.93370.86880.8765
Simplified EnAmDNN0.85660.81420.89470.85160.8576
Comparison of EnAmDNN and simplified EnAmDNN.

Conclusions

In this paper, we propose an ensemble deep learning framework (EnAmDNN) with an attention mechanism that aims to predict protein interaction networks. EnAmDNN firstly extracts the feature information of protein sequences through AC, LD, CT, and PseAAC, and projects the information into various feature spaces to segment information of AC, LD, CT, PseAAC amino acid from different perspective; then the multi-head attention mechanism is adopted to capture the internal connections of interactions; each technique is assigned 4 independent DNNs with different configurations, resulting in 16 basic learners, and finally combined by deep neural network. To further evaluate the prediction performance of EnAmDNN, we apply it to 5 independent data sets, where improvements of various degrees can be observed for indicators AUC, ACC, Recall, Precision, F1, from which it can be concluded that EnAmDNN learns better than previous approaches from different DNNs and representations.

Data Availability Statement

Alzheimer disease data was downloaded from the IntAct database (https://www.ebi.ac.uk/intact/query/annot:%22dataset:alzheimers%22?conversationContext=6) under search term annot: “dataset:alzheimers.” Cardiac data was downloaded from the IntAct database (https://www.ebi.ac.uk/intact/query/annot:%22dataset:cardiac%22?conversationContext=7) under search term annot: ‘dataset:cardiac.” Diabetes data was downloaded from the IntAct database (https://www.ebi.ac.uk/intact/query/annot:%22dataset:diabetes%22?conversationContext=8) under search term annot: “dataset:diabetes.” Parkinson's disease data was downloaded from the IntAct database (https://www.ebi.ac.uk/intact/query/annot:%22dataset:parkinsons%22?conversationContext=9) under search term annot: “dataset:parkinsons.” Cancer disease data was downloaded from the IntAct database (https://www.ebi.ac.uk/intact/query/annot:%22dataset:cancer%22?conversationContext=b) under search term annot: “dataset:cancer.” Protein sequence information were downloaded from the Uniprot database (https://www.uniprot.org/downloads).

Author Contributions

Conceived and designed the experiments, performed the experiments, and contributed reagents, materials, analysis tools: FL. Algorithm design and analysis and analyzed the data: FZ and FL. Wrote the paper: FL, FZ, XL, and QL.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Algorithm 1:

Basic learning algorithm based on multi-head attention with pooling

Input: interaction data D= { xi,yi } i=1n, encoded protein sequence pairxi = (si1, si2)
Return: basic learning algorithm parameters W, b

1: for i = 1 to n:

2: S1, S2xi(si1, si2) represent encoded protein p1 and protein p2

3: for iter = 1 to t:

4: S1, iter+1 = RELU (Batchβ, δ(Convλ(S1, iter)))

5: S2, iter+1 = RELU (Batchβ, δ(Convλ(S2, iter)))

6: end for

7: S1Convγ(S1, t)

8: S2Convγ(S2, t)

9: calculate the importance value of single attention by AttenheadS1i=softmax(S1S2dk)S1,AttenheadS2i=softmax(S2S1dk)S2

10: connect multiple attentions by S1*=MultiHead(S1,S2,S1),S2*=MultiHead(S2,S1,S2)

11: Calculate new feature importance value of multi-head attention with pooling by S1**=newMultiHead(S1*,S2*,S1*),S2**=newMultiHead(S2*,S1*,S2*)

12: merge the sequence feature pairs as the input of the network by Concat(S1**·S2**|S1**|×|S2**|,S1**AS2**,S1**,S2**)

13: output = softmax(W · merge + b)

14: end for

15: Return W, b

Algorithm 2:

EAM algorithm

Input: interaction data; basic learning algorithm ψ12,…,ψM; L layers; learning rate η; Max iterations and iteration threshold ε; class k ∈ {0, 1}
Return: ensemble strategy parameters W, b

1: initialize weight matrix W and Bias b, iter as 1

2: extract protein sequence from uniprot

3: vectorize protein sequence by AC, LD, CT, PseAAC

4: divide training data TrD={Di}i=1K={xi,yi}i=1n and testing data TeD

5: for m = 1 to M do

6: for k = 1 to K do

7: train hm=ϕm(D¯k) with D¯k

8: predict Dk and require predicted value Pmk = hm(Dk)

9: predict T and and require predicted value Tk = hm(T)

10: end for

11: splicing predicted value of training data through mth basic model by Pm=(Pm1,Pm2,,PmK)T

12: splicing predicted value of testing data through mth basic model by Tm = (T1 + T2 + … + TK)/K

13: end for

14: construct new training data by

15: and construct new testing data by newT = (T1, T2, …, TM)

16: Repeat:

17: iter + 1

18: for i = 1 to n:

19: set a(1)=xi*

20: for l=2 to L: al = σ(Wlal−1 + bl)

21: calculate probability pk divided into class k

22: calculate lossL=k=01ykL·log(pk)

22: for l=2 to L:

23: end for

24: for l=2 to L:

25: Wl=μi=1nlossl(al1)T,bl=μi=1nlossl

26: WlWl − ∇Wl, blbl − ∇bl

27: end for

27: Until ∇W, ∇b < ε

28: Return W, b

Table 3

PD with various methods.

MethodsAccuracyRecallPrecisionF1AUC
EnsAmDNN0.8950.85680.92750.89030.8951
SVM_AC0.7970.620.93730.74440.7904
kNN_LD0.80570.80420.79060.79580.8064
SVM_LD0.83150.73370.90840.81080.831
NDDs_APAAC0.87730.82470.90580.86270.8906
EnsDNN0.89170.84330.93110.88460.8915

The highest score of each evaluation criteria is emphasized in bold.

Table 4

Cancer with various methods.

MethodsAccuracyRecallPrecisionF1AUC
EnsAmDNN0.85020.80620.88630.84360.8508
SVM_AC0.65240.48480.73470.58110.6545
kNN_LD0.64750.67610.62020.64580.6471
SVM_LD0.66730.55910.7150.62630.6681
NDDs_APAAC0.75510.72240.77640.74740.7555
EnsDNN0.80080.75490.83620.79250.802

The highest score of each evaluation criteria is emphasized in bold.

Table 5

Cancer with various methods.

MethodsAccuracyRecallPrecisionF1AUC
EnsAmDNN0.9070.85230.960.90180.9088
SVM_AC0.73560.63160.8060.70370.7371
kNN_LD0.76710.72460.79340.75650.7675
SVM_LD0.78190.75450.79620.7740.7816
NDDs_APAAC0.84540.83260.8370.83390.8446
EnsDNN0.90390.87470.92520.8990.9034

The highest score of each evaluation criteria is emphasized in bold.

  28 in total

1.  DeepPPI: Boosting Prediction of Protein-Protein Interactions with Deep Neural Networks.

Authors:  Xiuquan Du; Shiwei Sun; Changlin Hu; Yu Yao; Yuanting Yan; Yanping Zhang
Journal:  J Chem Inf Model       Date:  2017-05-26       Impact factor: 4.956

2.  Chemical-protein interaction extraction via contextualized word representations and multihead attention.

Authors:  Yijia Zhang; Hongfei Lin; Zhihao Yang; Jian Wang; Yuanyuan Sun
Journal:  Database (Oxford)       Date:  2019-01-01       Impact factor: 3.451

3.  MusiteDeep: a deep-learning framework for general and kinase-specific phosphorylation site prediction.

Authors:  Duolin Wang; Shuai Zeng; Chunhui Xu; Wangren Qiu; Yanchun Liang; Trupti Joshi; Dong Xu
Journal:  Bioinformatics       Date:  2017-12-15       Impact factor: 6.937

4.  Structure-based prediction of protein-protein interactions on a genome-wide scale.

Authors:  Qiangfeng Cliff Zhang; Donald Petrey; Lei Deng; Li Qiang; Yu Shi; Chan Aye Thu; Brygida Bisikirska; Celine Lefebvre; Domenico Accili; Tony Hunter; Tom Maniatis; Andrea Califano; Barry Honig
Journal:  Nature       Date:  2012-09-30       Impact factor: 49.962

5.  IntAct--open source resource for molecular interaction data.

Authors:  S Kerrien; Y Alam-Faruque; B Aranda; I Bancarz; A Bridge; C Derow; E Dimmer; M Feuermann; A Friedrichsen; R Huntley; C Kohler; J Khadake; C Leroy; A Liban; C Lieftink; L Montecchi-Palazzi; S Orchard; J Risse; K Robbe; B Roechert; D Thorneycroft; Y Zhang; R Apweiler; H Hermjakob
Journal:  Nucleic Acids Res       Date:  2006-12-01       Impact factor: 16.971

6.  The Universal Protein Resource (UniProt).

Authors:  Amos Bairoch; Rolf Apweiler; Cathy H Wu; Winona C Barker; Brigitte Boeckmann; Serenella Ferro; Elisabeth Gasteiger; Hongzhan Huang; Rodrigo Lopez; Michele Magrane; Maria J Martin; Darren A Natale; Claire O'Donovan; Nicole Redaschi; Lai-Su L Yeh
Journal:  Nucleic Acids Res       Date:  2005-01-01       Impact factor: 16.971

7.  Sequence-based prediction of protein protein interaction using a deep-learning algorithm.

Authors:  Tanlin Sun; Bo Zhou; Luhua Lai; Jianfeng Pei
Journal:  BMC Bioinformatics       Date:  2017-05-25       Impact factor: 3.169

8.  Dual Convolutional Neural Networks With Attention Mechanisms Based Method for Predicting Disease-Related lncRNA Genes.

Authors:  Ping Xuan; Yangkun Cao; Tiangang Zhang; Rui Kong; Zhaogong Zhang
Journal:  Front Genet       Date:  2019-05-03       Impact factor: 4.599

9.  Deep learning of the tissue-regulated splicing code.

Authors:  Michael K K Leung; Hui Yuan Xiong; Leo J Lee; Brendan J Frey
Journal:  Bioinformatics       Date:  2014-06-15       Impact factor: 6.937

10.  De novo identification of replication-timing domains in the human genome by deep learning.

Authors:  Feng Liu; Chao Ren; Hao Li; Pingkun Zhou; Xiaochen Bo; Wenjie Shu
Journal:  Bioinformatics       Date:  2015-11-05       Impact factor: 6.937

View more
  8 in total

1.  Identification of key genes associated with sepsis patients infected by staphylococcus aureus through weighted gene co-expression network analysis.

Authors:  Han Wu; Haoting Chen; Junjie Wang; Shaohua Yin; Jiaqian Huang; Zhiqiang Wang; Xiaojie Zhang; Minghua Wang
Journal:  Am J Transl Res       Date:  2021-12-15       Impact factor: 4.060

2.  An Augmented High-Dimensional Graphical Lasso Method to Incorporate Prior Biological Knowledge for Global Network Learning.

Authors:  Yonghua Zhuang; Fuyong Xing; Debashis Ghosh; Farnoush Banaei-Kashani; Russell P Bowler; Katerina Kechris
Journal:  Front Genet       Date:  2022-01-27       Impact factor: 4.599

Review 3.  Protein Function Analysis through Machine Learning.

Authors:  Chris Avery; John Patterson; Tyler Grear; Theodore Frater; Donald J Jacobs
Journal:  Biomolecules       Date:  2022-09-06

4.  Molecular Modelling Hurdle in the Next-Generation Sequencing Era.

Authors:  Guerau Fernandez; Dèlia Yubero; Francesc Palau; Judith Armstrong
Journal:  Int J Mol Sci       Date:  2022-06-28       Impact factor: 6.208

5.  Benchmark Evaluation of Protein-Protein Interaction Prediction Algorithms.

Authors:  Brandan Dunham; Madhavi K Ganapathiraju
Journal:  Molecules       Date:  2021-12-22       Impact factor: 4.927

6.  SDNN-PPI: self-attention with deep neural network effect on protein-protein interaction prediction.

Authors:  Xue Li; Peifu Han; Gan Wang; Wenqi Chen; Shuang Wang; Tao Song
Journal:  BMC Genomics       Date:  2022-06-27       Impact factor: 4.547

Review 7.  Overview of methods for characterization and visualization of a protein-protein interaction network in a multi-omics integration context.

Authors:  Vivian Robin; Antoine Bodein; Marie-Pier Scott-Boyer; Mickaël Leclercq; Olivier Périn; Arnaud Droit
Journal:  Front Mol Biosci       Date:  2022-09-08

8.  AutoPPI: An Ensemble of Deep Autoencoders for Protein-Protein Interaction Prediction.

Authors:  Gabriela Czibula; Alexandra-Ioana Albu; Maria Iuliana Bocicor; Camelia Chira
Journal:  Entropy (Basel)       Date:  2021-05-21       Impact factor: 2.524

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.