Literature DB >> 31510705

Multifaceted protein-protein interaction prediction based on Siamese residual RCNN.

Muhao Chen1, Chelsea J-T Ju1, Guangyu Zhou1, Xuelu Chen1, Tianran Zhang2, Kai-Wei Chang1, Carlo Zaniolo1, Wei Wang1.   

Abstract

MOTIVATION: Sequence-based protein-protein interaction (PPI) prediction represents a fundamental computational biology problem. To address this problem, extensive research efforts have been made to extract predefined features from the sequences. Based on these features, statistical algorithms are learned to classify the PPIs. However, such explicit features are usually costly to extract, and typically have limited coverage on the PPI information.
RESULTS: We present an end-to-end framework, PIPR (Protein-Protein Interaction Prediction Based on Siamese Residual RCNN), for PPI predictions using only the protein sequences. PIPR incorporates a deep residual recurrent convolutional neural network in the Siamese architecture, which leverages both robust local features and contextualized information, which are significant for capturing the mutual influence of proteins sequences. PIPR relieves the data pre-processing efforts that are required by other systems, and generalizes well to different application scenarios. Experimental evaluations show that PIPR outperforms various state-of-the-art systems on the binary PPI prediction problem. Moreover, it shows a promising performance on more challenging problems of interaction type prediction and binding affinity estimation, where existing approaches fall short.
AVAILABILITY AND IMPLEMENTATION: The implementation is available at https://github.com/muhaochen/seq_ppi.git. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
© The Author(s) 2019. Published by Oxford University Press.

Entities:  

Mesh:

Substances:

Year:  2019        PMID: 31510705      PMCID: PMC6681469          DOI: 10.1093/bioinformatics/btz328

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.937


1 Introduction

Detecting protein–protein interactions (PPIs) and characterizing the interaction types are essential toward understanding cellular biological processes in normal and disease states. Knowledge from these studies potentially facilitates therapeutic target identification (Petta ) and novel drug design (Skrabanek ). High-throughput experimental technologies have been rapidly developed to discover and validate PPIs on a large scale. These technologies include yeast two-hybrid screens (Fields and Song, 1989), tandem affinity purification (Gavin ) and mass spectrometric protein complex identification (Ho ). However, experiment-based methods remain expensive, labor-intensive and time-consuming. Most importantly, they often suffer from high levels of false-positive predictions (Sun ; You ). Evidently, there is an immense need for reliable computational approaches to identify and characterize PPIs. The amino acid sequence represents the primary structure of a protein, which is the simplest type of information either obtained through direct sequencing or translated from DNA sequences. Many research efforts address the PPI problem based on predefined features extracted from protein sequences, such as ontological features of amino acids (Jansen ), autocovariance (AC) (Guo ), conjoint triads (CT) (Shen ) and composition-transition-distribution (CTD) descriptors (Yang ). These features generally summarize specific aspects of protein sequences such as physicochemical properties, frequencies of local patterns and the positional distribution of amino acids. On top of these features, several statistical learning algorithms (Guo ; Huang ; You , 2015) are applied to predict PPIs in the form of binary classification. These approaches provide feasible solutions to the problem. However, the extracted features used in these approaches only have limited coverage on interaction information, as they are dedicated to specific facets of the protein profiles. To mitigate the inadequacy of statistical learning methods, deep learning algorithms provide the powerful functionality to process large-scale data and automatically extract useful features for objective tasks (LeCun ). Recently, deep learning architectures have produced powerful systems to address several bioinformatics problems related to single nucleotide sequences, such as genetic variants detection (Anderson, 2018), DNA function classification (Quang and Xie, 2016), RNA-binding site prediction (Zhang ) and chromatin accessibility prediction (Min ). These works typically use convolutional neural networks (CNN) (Anderson, 2018; Zhang ) for automatically selecting local features, or recurrent neural networks (RNN) (Quang and Xie, 2016) that aim at preserving the contextualized and long-term ordering information. In contrast, fewer efforts (discussed in Section 2) have been made to capture the pairwise interactions of proteins with deep learning, which remains a non-trivial problem with the following challenges: (i) Characterization of the proteins requires a model to effectively filter and aggregate their local features, while preserving significant contextualized and sequential information of the amino acids; (ii) extending a deep neural architecture often leads to inefficient learning processes, and suffers from the notorious vanishing gradient problem (Pascanu ); (iii) an effective mechanism is also needed to apprehend the mutual influence of protein pairs in PPI prediction. Moreover, it is essential for the framework to be scalable to large data, and to be generalized to different prediction tasks. In this paper, we introduce PIPR (Protein–Protein Interaction Prediction Based on Siamese Residual RCNN), a deep learning framework for PPI prediction using only the sequences of a protein pair. PIPR employs a Siamese architecture to capture the mutual influence of a protein sequence pair. The learning architecture is based on a residual recurrent convolutional neural network (RCNN), which integrates multiple occurrences of convolution layers and residual gated recurrent units. To represent each amino acid in this architecture, PIPR applies an efficient property-aware lexicon embedding approach to better capture the contextual and physicochemical relatedness of amino acids. This comprehensive encoding architecture provides a multi-granular feature aggregation process to effectively leverage both sequential and robust local information of the protein sequences. It is important to note that the scope of this work focuses only on the primary sequence as it is the fundamental information to describe a protein. Our contributions are 3-fold. First, we construct an end-to-end framework for PPI prediction that relieves the data pre-processing efforts for users. PIPR requires only the primary protein sequences as the input, and is trained to automatically preserve the critical features from the sequences. Second, we emphasize and demonstrate the needs of considering the contextualized and sequential information when modeling the PPIs. Third, the architecture of PIPR can be flexibly used to address different PPI tasks. Besides the binary prediction that is widely attempted in previous works, our framework extends its use to two additional challenging problems: multi-class interaction type prediction and binding affinity estimation. We use five datasets to evaluate the performance of our framework on these tasks. PIPR outperforms various state-of-the-art approaches on the binary prediction task, which confirms the effectiveness in terms of integrating both local features and sequential information. The promising performance of the other two tasks demonstrates the wide usability of our approach. Especially on the binding affinity estimation of mutated proteins, PIPR is able to respond to the subtle changes of point mutations and provides the best estimation with the smallest errors.

2 Related works

Sequence-based approaches provide a critical solution to the binary PPI prediction task. Homology-based methods (Philipp ) rely on BLAST to map a pair of sequences to known interacting proteins. Alternatively, other works address the task with statistical learning models, including SVM (Guo ; You ), kNN (Yang ), Random Forest (Wong ), multi-layer perceptron (MLP) (Du ) and ensemble ELM (EELM) (You ). These approaches rely on several feature extraction processes for the protein sequences, such as CT (Sun ; You ), AC (Guo ; Sun ; You ), CTD (Du ; Yang ), multi-scale continuous and discontinuous (MCD) descriptors (You ) and local phase quantization (LPQ) (Wong ). These features measure physicochemical properties of the 20 canonical amino acids, and aim at summarizing full sequence information relevant to PPIs. More recent works (Sun ; Wang ) propose the use of stacked autoencoders (SAE) to refine these heterogeneous features in low-dimensional spaces, which improve the aforementioned models on the binary prediction task. On the contrary, fewer efforts have been made toward multi-class prediction to infer the interaction types (Silberberg ; Zhu ) and the regression task to estimate binding affinity (Srinivasulu ; Yugandhar and Gromiha, 2014). These methods have largely relied on their capability of extracting and selecting better features, while the extracted features are far from fully exploiting the interaction information. By nature, the PPI prediction task is comparable to the neural sentence pair modeling tasks in natural language processing (NLP) research, as they both seek to characterize the mutual influence of two sequences based on their latent features. In NLP, neural sentence pair models typically focus on capturing the discourse relations of lexicon sequences, such as textual entailment (Hu ; Yin ), paraphrases (He ; Yin and Schütze, 2015) and sub-topic relations (Chen ). Many recent efforts adopt a Siamese encoding architecture, where encoders based on CNN (Hu ; Yin and Schütze, 2015) and RNN (Mueller and Thyagarajan, 2016) are widely used. A binary classifier is then stacked to the sequence pair encoder for the detection of a discourse relation. In contrast to sentences, proteins are profiled in sequences with more intractable patterns, as well as in a drastically larger range of lengths. Precisely capturing the PPI requires much more comprehensive learning architectures to distill the latent information from the entire sequences, and to preserve the long-term ordering information. One recent work (Hashemifar ), DPPI, uses a deep CNN-based architecture which focuses on capturing local features from protein profiles. DPPI represents the first work to deploy deep learning to PPI prediction, and has achieved the state-of-the-art performance on the binary prediction task. However, it requires excessive efforts for data pre-processing such as constructing protein profiles by PSI-BLAST (Altschul ), and does not incorporate a neural learning architecture that captures the important contextualized and sequential features. DNN-PPI (Li ) represents another relevant work of this line, which deploys a different learning structure with two separated CNN encoders. However, DNN-PPI does not incorporate physicochemical properties into amino acid representations, and does not employ a Siamese learning architecture to fully characterize pairwise relations of sequences.

3 Materials and methods

We introduce an end-to-end deep learning framework, PIPR, for sequence-based PPI prediction tasks. The overall learning architecture is illustrated in Figure 1. PIPR employs a Siamese architecture of residual RCNN encoder to better apprehend and utilize the mutual influence of two sequences. To capture the features of the protein sequences from scratch, PIPR pre-trains the embeddings of canonical amino acids to capture their contextual similarity and physicochemical properties. The latent representation of each protein in a protein pair is obtained by feeding the corresponding amino acid embeddings into the sequence encoder. The embeddings of these two sequences are then combined to form a sequence pair vector. Finally, this sequence pair vector is fed into an MLP with appropriate loss functions, suiting for specific prediction tasks. In this section, we describe the details of each model component. We begin with the denotations and problem specifications.
Fig. 1.

The overall learning architecture of our framework

The overall learning architecture of our framework

3.1 Preliminary

We use A to denote the vocabulary of 20 canonical amino acids. A protein is profiled as a sequence of amino acids such that each. For each amino acid a, we use bold-faced to denote its embedding representation, which we are going to specify in Section 3.2.3. We use I to denote the set of protein pairs, and denotes a pair of proteins of which our framework captures the interaction. We address three challenging PPI prediction tasks based only on the primary sequence information: (i) Binary prediction seeks to provide a binary classifier to indicate whether the corresponding protein pair interacts, which is the simplest and widely considered problem setting in previous works (Hashemifar ; Skrabanek ; Sun ). (ii) Interaction type prediction is a multi-class classification problem, which seeks to identify the interaction type of two proteins. (iii) Binding affinity estimation aims at producing a regression model to estimate the strength of the binding interaction.

3.2 RCNN-based protein sequence encoder

We employ a deep Siamese architecture of residual RCNN to capture latent semantic features of the protein sequence pairs.

Residual RCNN

The RCNN seeks to leverage both the global sequential information and local features that are significant to the characterization of PPI from the protein sequences. This deep neural encoder stacks multiple instances of two computational modules, i.e. convolution layers with pooling and bidirectional residual gated recurrent units. The architecture of an RCNN unit is shown on the left of Figure 2.
Fig. 2.

The structure of our residual RCNN encoder is shown on the right, and the RCNN unit is shown on the left. Each RCNN unit contains a convolution-pooling layer followed a bidirectional residual GRU

The structure of our residual RCNN encoder is shown on the right, and the RCNN unit is shown on the left. Each RCNN unit contains a convolution-pooling layer followed a bidirectional residual GRU The convolution layer with pooling. We use to denote an input vector sequence that corresponds to the embedded amino acids or the outputs of a previous neural layer. A convolution layer applies a weight-sharing kernel to generate a k-dimension latent vector from a window of the input vector sequence X: for which h is the kernel size, and is a bias vector. The convolution layer applies the kernel as a sliding window to produce a sequence of latent vectors , where each latent vector combines the local features from each h-gram of the input sequence. The n-max-pooling mechanism is applied to every consecutive n-length subsequence (i.e. non-overlapped n-strides) of the convolution outputs, which takes the maximum value along each dimension j by . The purpose of this mechanism is to discretize the convolution results, and preserve the most significant features within each n-stride (Chen ; Hashemifar ; Kim, 2014). By definition, this mechanism divides the size of processed features by n. The outputs from the max-pooling are fed into the bidirectional gated recurrent units in our RCNN encoder. The residual gated recurrent units. The gated recurrent unit model (GRU) represents an alternative to the long-short-term memory (LSTM) network (Cho ), which consecutively characterizes the sequential information without using separated memory cells (Dhingra ). Each unit consists of two types of gates to track the state of the sequence, i.e. the reset gate and the update gate. Given the embedding of an incoming item (either a pre-trained amino acid embedding, or an output of the previous layer), GRU updates the hidden state of the sequence as a linear combination of the previous state and the candidate state of a new item , which is calculated as below. Notation thereof denotes the element-wise multiplication. The update gate balances the information of the previous sequence and the new item, where capitalized and denote different weight matrices, denote bias vectors and σ is the sigmoid function. The candidate state is calculated similarly to those in a traditional recurrent unit, and the reset gate controls how much information of the past sequence contributes to . Note that GRU generally performs comparably to LSTM in sequence encoding tasks, but is less complex and requires much fewer computational resources for training. A bidirectional GRU layer characterizes the sequential information in two directions. It contains the forward encoding process that reads the input vector sequence from to , and a backward encoding process that reads in the opposite direction. The encoding results of both processes are concatenated for each input item, i.e. . The residual mechanism passes on an identity mapping of the GRU inputs to its output side through a residual shortcut (He ). By adding the forwarded input values to the outputs, the corresponding neural layer is only required to capture the difference between the input and output values. This mechanism aims at improving the learning process of non-linear neural layers by increasing the sensitivity of the optimization gradients (He ; Kim ), as well as preventing the model from the vanishing gradient problem. It has been widely deployed in deep learning architectures for various tasks of image recognition (He ), document classification (Conneau ) and speech recognition (Zhang ). In our deep RCNN, the bidirectional GRU is incorporated with the residual mechanism, and will pass on the following outputs to its subsequent neural network layer: In our development, we have found that the residual mechanism is able to drastically simplify the training process, and largely decreases the epochs of parameter updates for the model to converge.

Protein sequence encoding

Figure 2 shows the entire structure of our RCNN encoder. The RCNN encoder alternately stacks multiple occurrences of the above two intermediary neural network components. A convolution layer serves as the first encoding layer to extract local features from the input sequence. On top of that, a residual GRU layer takes in the preserved local features, whose outputs are passed to another convolution layer. Repeating of these two components in the network structure conducts an automatic multi-granular feature aggregation process on the protein sequence, while preserving the sequential and contextualized information on each granularity of the selected features. The last residual GRU layer is followed by another convolution layer for a final round of local feature selection to produce the last hidden states . Note that the dimensionality of the last hidden states does not need to equal that of the previous hidden states. A high-level sequence embedding of the entire protein sequence is obtained from the global average-pooling (Lin ) of, i.e..

Pre-trained amino acid embeddings

To support inputting the non-numerical sequence information, we provide a useful embedding method to represent each amino acid as a semi-latent vector a. Each embedding vector is a concatenation of two sub-embeddings, i.e.. The first part measures the co-occurrence similarity of the amino acids, which is obtained by pre-training the Skip-Gram model (Mikolov ) on protein sequences. The learning objective of Skip-Gram is to minimize the following negative log likelihood loss. thereof is the first-part embedding of the t-th amino acid is that of a neighboring amino acid, and C is the size of half context. (The context of Skip-Gram means a subsequence of a given protein sequence S, such that the subsequence is of length.) The probability p is defined as the following softmax: where n is the negative sampling size, and is a negative sample that does not co-occur with in the same context. The second part represents the similarity of electrostaticity and hydrophobicity among amino acids. The 20 amino acids can be clustered into 7 classes based on their dipoles and volumes of the side chains to reflect this property. Thus, is a one-hot encoding based on the classification defined by Shen .

3.3 Learning architecture and learning objectives

Our framework characterizes the interactions in the following two stages.

Siamese architecture

Given a pair of proteins , the same RCNN encoder is used to obtain the sequence embeddings and of both proteins. Both sequence embeddings are combined using element-wise multiplication, i.e.. This is a commonly used operation to infer the relation of sequence embeddings (Hashemifar ; Jiang ; Rocktäschel ; Tai ). Note that some works use the concatenation of sequence embeddings (Sun ; Yin and Schütze, 2015) instead of multiplication, which we find to be less effective in modeling the symmetric relations of proteins.

Learning objectives

An MLP with leaky ReLU (Maas ) is applied to the previous sequence pair representation, whose output is either a vector or a scalar, depending on whether the model solves a classification or a regression task for the protein pair p. The entire learning architecture is trained to optimize the following two types of losses according to different PPI prediction problems. Cross-entropy loss is optimized for the two classification problems, i.e. binary prediction and interaction type prediction. In this case, the MLP output is a vector, whose dimensionality equals the number of classes m. is normalized by a softmax function, where the i-th dimension corresponds to the confidence score for the i-th class. The learning objective is to minimize the following cross-entropy loss, where c is a one-hot indicator for the class label of protein pair p. Mean squared loss is optimized for the binding affinity estimation task. In this case, is a scalar output that is normalized by a sigmoid function, which is trained to approach the normalized ground truth score by minimizing the following objective function:

4 Experiments

We present the experimental evaluation of the proposed framework on three PPI prediction tasks, i.e. binary prediction, multi-class interaction type prediction and binding affinity estimation. The experiments are conducted on the following datasets.

4.1 Datasets

Guo’s datasets. Guo generate several datasets from different species for the binary prediction of PPIs. Each dataset contains a balanced number of positive and negative samples. Among these resources, the Yeast dataset is a widely used benchmark by most state-of-the-art methods (Hashemifar ; Wong ; You , 2014). There are 2497 proteins forming 11 188 cases of PPIs, with half of them representing the positive cases, and the other half the negative cases. The positive cases are selected from the database of interacting proteins DIP_20070219 (Salwinski ), where proteins with fewer than 50 amino acids or ≥40% sequence identity are excluded. We use the full protein sequences in our model, which are obtained from the UniProt (Consortium ). The negative cases are generated by randomly pairing the proteins without evidence of interaction, and filtered by their sub-cellular locations. In other words, non-interactive pairs residing in the same location are excluded. In addition, we combine the data for Caenorhabditis elegans, Escherichia coli and Drosophila melanogaster as the multi-species dataset. We use the cluster analysis of the CD-HIT (Li and Godzik, 2006) program to generate non-redundant subsets. Proteins with fewer than 50 amino acids or high sequence identify (40, 25, 10 or 1%) are removed. STRING datasets. The STRING database (Szklarczyk ) annotates PPIs with their types. There are seven types of interactions: activation, binding, catalysis, expression, inhibition, post-translational modification (ptmod) and reaction. We download all interaction pairs for Homo sapiens from database version 10.5 (Szklarczyk ), along with their full protein sequences. Among the corresponding proteins, we randomly select 3000 proteins and 8000 proteins that share <40% of sequence identity to generate two subsets. In this process, we randomly sample instances of different interaction types to ensure a balanced class distribution. Eventually, the two generated datasets, denoted by SHS27k and SHS148k, contain 26 945 cases and 148 051 cases of interactions respectively. We use these two datasets for the PPI type prediction task. SKEMPI dataset. We obtain the protein binding affinity data from SKEMPI (the structural database of kinetics and energetics of mutant protein interactions) (Moal and Fernández-Recio, 2012) for the affinity estimation task. It contains 3047 binding affinity changes upon mutation of protein sub-units within a protein complex. The binding affinity is measured by equilibrium dissociation constant (K), reflecting the strength of biomolecular interactions. The smaller K value means the higher binding affinity. Each protein complex contains single or multiple amino acid substitutions. The sequence of the protein complex is retrieved from the protein data bank (PDB) (Berman ). We manually replace the mutated amino acids. For duplicate entries, we take the average K. The final dataset results in the binding affinity of 2792 mutant protein complexes, along with 158 wild-types.

4.2 Binary PPI prediction

Binary PPI prediction is the primary task targeted by a handful of previous works (Hashemifar ; Shen ; Sun ; Yang ; You ). The objective of these works is to identify whether a given pair of proteins interacts or not based on their sequences. We evaluate PIPR based on Guo’s datasets. The Yeast benchmark dataset thereof is used to compare PIPR with various baseline approaches, and the multi-species dataset is to demonstrate PIPR’s capability of predicting interactions for proteins of different species that share very low sequence identity with those in training. The baseline approaches include SVM-AC (Guo ), kNN-CTD (Yang ), EELM-PCA (You ), SVM-MCD (You ), MLP (Du ), Random Forest LPQ (RF-LPQ) (Wong ), SAE (Sun ), DNN-PPI (Li ) and DPPI (Hashemifar ). In addition, we report the results of a Siamese Residual GRU (SRGRU) architecture, which is a simplification of PIPR, where we discard all intermediary convolution layers and keep only the bidirectional residual GRU. The purpose of SRGRU is to show the significance of the contextualized and sequential information of protein profiles in characterizing PPIs. We also report the results of Siamese CNN (SCNN) by removing the residual GRU in PIPR. This degenerates our framework to a similar architecture to DPPI, but differs in that SCNN directly conducts an end-to-end training on raw sequences instead of requiring the protein profiles constructed by PSI-BLAST. We use AMSGrad (Reddi ) to optimize the cross-entropy loss, for which we set the learning rate α to 0.001, the exponential decay rates β1 and β2 to 0.9 and 0.999, and batch size to 256 on both datasets. The number of occurrences for the RCNN units (i.e. one convolution-pooling layer followed by one bidirectional residual GRU layer) is set to five, where we adopt three-max-pooling and the convolution kernel of size three. We set the hidden state size to be 50, and the RCNN output size to be 100. We set this configuration to ensure the RCNN to compress the selected features in a reasonably small vector sequence, before the features are aggregated by the last global average-pooling. We zero-pad short sequences to the longest sequence length in the dataset. This is a widely adopted technique for sequence modeling in NLP (Chen ; He ; Hu ; Yin ; Zhou ) as well as in bioinformatics (Min ; Müller ; Pan and Shen, 2018) for efficient training. Note that the configuration of embedding pre-training is discussed in Section 4.5, and the model configuration study of different hyperparameter values is provided in the Supplementary Material. All model variants are trained until converge at each fold of the cross-validation (CV). Evaluation protocol. Following the settings in previous works (Hashemifar ; Shen ; Sun ; You , 2015), we conduct 5-fold CV on the Yeast dataset. Under the k-fold CV setting, the data are equally divided into k non-overlapping subsets, and each subset has a chance to train and to test the model so as to ensure an unbiased evaluation. We aggregate fix metrics on the test cases of each fold, i.e. the overall accuracy, precision, sensitivity, specificity, F1 and Matthews correlation coefficient (MCC) on positive cases. All these metrics are preferred to be higher to indicate better performance. Based on the reported accuracy over 5-folds, we also conduct two-tailed Welch’s t-tests (Welch, 1947) to evaluate the significance of the improvement on different pairs of approaches. The P-values are adjusted by the Benjamini–Hochberg procedure (Benjamini and Hochberg, 1995) to control the false discovery rate for multiple hypothesis testing. Results. As shown in Table 1, the CNN-based architecture, DPPI, demonstrates state-of-the-art performance over other baselines that employ statistical learning algorithms or densely connected MLP (We are unable to obtain the source codes of two deep-learning methods, SAE and DNN-PPI. We implement these two models following the descriptions in their papers. Our implementations are verified by achieving comparable performance on the Pan’s dataset (Pan ) as reported in the papers. However, these two implementations can only achieve 67.17 and 76.61% in overall accuracy respectively on the Yeast dataset.). This shows the superiority of deep-learning-based techniques in encapsulating various types of information of a protein pair, such as amino acid composition and their co-occurrences, and automatically extracting the robust ones for the learning objectives. That said, DPPI requires an extensive effort in data pre-processing, specifically in constructing the protein profile for each sequence. On average, each PSI-BLAST search of a protein against the NCBI non-redundant protein database (184 243 125 sequences) requires around 90 min of computation on our server. Even with eight cores, each search finishes in 15 min. We estimate that processing 2497 sequences of the Yeast dataset from scratch can take about 26 days. It is worth mentioning that PIPR only requires 8 s to pre-train the amino acid embedding, and 2.5 min to train on the Yeast dataset (Table 7). We implement SCNN to evaluate the performance of a simplified CNN architecture, which produces comparable results as DPPI. These two frameworks show that CNN can already leverage the significant features from primary protein sequences.
Table 1.

Evaluation of binary PPI prediction on the Yeast dataset based on 5-fold cross-validation. We report the mean and SD for the test sets

MethodsAccuracy (%)Precision (%)Sensitivity (%)Specificity (%)F1-score (%)MCC (%)
SVM-AC87.35 ± 1.3887.82 ± 4.8487.30 ± 5.2387.41 ± 6.3387.34 ± 1.3375.09 ± 2.51
kNN-CTD86.15 ± 1.1790.24 ± 1.3481.03 ± 1.74NA85.39 ± 1.51NA
EELM-PCA86.99 ± 0.2987.59 ± 0.3286.15 ± 0.43NA86.86 ± 0.3777.36 ± 0.44
SVM-MCD91.36 ± 0.491.94 ± 0.6990.67 ± 0.77NA91.3 ± 0.7384.21 ± 0.66
MLP94.43 ± 0.396.65 ± 0.5992.06 ± 0.36NA94.3 ± 0.4588.97 ± 0.62
RF-LPQ93.92 ± 0.3696.45 ± 0.4591.10 ± 0.31NA93.7 ± 0.3788.56 ± 0.63
SAE67.17 ± 0.6266.90 ± 1.4268.06 ± 2.5066.30 ± 2.2767.44 ± 1.0834.39 ± 1.25
DNN-PPI76.61 ± 0.5175.1 ± 0.6679.63 ± 1.3473.59 ± 1.2877.29 ± 0.6653.32 ± 1.05
DPPI94.5596.6892.24NA94.41NA
SRGRU93.77 ± 0.8494.60 ± 0.6492.85 ± 1.5894.69 ± 0.8193.71 ± 0.8587.56 ± 1.67
SCNN95.03 ± 0.4795.51 ± 0.7794.51 ± 1.2795.55 ± 0.7795.00 ± 0.5090.08 ± 0.93
PIPR 97.09 ± 0.24 97.00 ± 0.65 97.17 ± 0.4497.00 ± 0.67 97.09 ± 0.23 94.17 ± 0.48

Each boldfaced number indicates the best of the corresponding metric.

NA, not available from the original paper.

Table 7.

Run-time of training embeddings and different prediction tasks

TaskEmbeddingsBinaryMulti-classMulti-classRegression
DatasetSHS148kYeastSHS27kSHS148kSKEMPI
Sample size800011 18826 945148 0512 950
Training time8 s2.5 min15.8 min138.3 min12.5 min
Evaluation of binary PPI prediction on the Yeast dataset based on 5-fold cross-validation. We report the mean and SD for the test sets Each boldfaced number indicates the best of the corresponding metric. NA, not available from the original paper. In addition, the SRGRU architecture has offered comparable performance to SCNN. This indicates that preserving the sequential and contextualized features of the protein sequences is as crucial as incorporating the local features. By integrating both significant local features and sequential information, PIPR outperforms DPPI by 2.54% in accuracy, 4.93% in sensitivity and 2.68% in F1-score. Next, we evaluate whether the improved accuracy of PIPR is statistically significant. Table 2 reports the P-values of SRGRU, SCNN and PIPR compared to other baseline approaches, where the statistically significant comparisons (P-values <0.01) are highlighted in red. Since the SD of DPPI is unavailable, we are not able to include DPPI in this analysis. The evaluation shows that PIPR performs statistically significantly better than all other approaches, including SCNN and SRGRU. On the other hand, SCNN is not statistically significantly better than SRGRU. Thus, the residual RCNN is very promising for modeling binary PPIs.
Table 2.

Statistical assessment (t-test; two-tailed) on the accuracy of binary PPI prediction

P-valueSRGRUSCNN PIPR
SVM-AC9.69E-051.22E-049.69E-05
kNN-CTD1.03E-052.23E-052.84E-05
EELM-PCA2.33E-053.94E-082.43E-10
SVM-MCD1.67E-032.60E-061.35E-07
MLP1.71E-015.29E-021.12E-06
RF-LPQ7.28E-014.10E-031.75E-06
SAE4.27E-101.78E-104.19E-09
DNN-PPI1.62E-082.27E-102.70E-09
SRGRUNA2.87E-026.60E-04
SCNN2.87E-02NA1.80E-04

Note: The statistically significant differences are highlighted in red.

NA, not available.

Statistical assessment (t-test; two-tailed) on the accuracy of binary PPI prediction Note: The statistically significant differences are highlighted in red. NA, not available. We also report the 5-fold CV performance of PIPR on variants of the multi-species dataset, where proteins are excluded based on different thresholds of sequence identity. The results in Table 3 show that PIPR performs consistently well under lenient and stringent criteria of sequence identity between training and testing. More importantly, PIPR is able to train and test on multiple species, and is robust against extremely low sequence identity of <1%.
Table 3.

Evaluation of binary PPI prediction on variants of multi-species (C. elegans, D. melanogaster and E. coli) dataset

Seq. identity# of proteinsPos. pairsNeg. pairsAccuracy (%)F1-score (%)
Any11 52932 95932 95998.1998.17
<0.40973925 91622 01298.2998.28
<0.25779019 45815 82797.9198.08
<0.10576912 641981997.5497.79
<0.01517110 747806597.5197.80
Evaluation of binary PPI prediction on variants of multi-species (C. elegans, D. melanogaster and E. coli) dataset

4.3 Interaction type prediction

The objective of this task is to predict the interaction type of two interacting proteins. We evaluate this task based on SHS27k and SHS148k datasets. To the best of our knowledge, much fewer efforts attempt for the multi-class PPI prediction in contrast to the binary prediction. Zhu train a two-stage SVM classifier to distinguish obligate, non-obligate and crystal packing interactions; Silberberg use logistic regression to predict several types of enzymatic actions. However, none of their implementations is publicly available. Different from the categories of interaction types used above, we aim at predicting the interaction types annotated by the STRING database. We train several statistical learning algorithms on the widely employed AC and CTD features for protein characterization as our baselines. These algorithms include SVM, Random Forest, AdaBoost [SAMME.R algorithm (Zhu )], kNN classifier and logistic regression. For deep-learning-based approaches, we deploy the SCNN architecture where an output MLP with categorical cross-entropy loss is incorporated, as well as a similar SRGRU architecture into comparison. Results of two naïve baselines of random guessing and zero rule (i.e. simply predicting the majority class) are also reported for reference. Evaluation protocol. All approaches are evaluated on the two datasets by 10-fold CV, using the same partition scheme for a more unbiased evaluation (James ; McLachlan ). We carry forward the model configurations from the last experiment to evaluate the performance of the frameworks under controlled variables. For baseline models, we examine three different ways of combining the feature vectors of the two input proteins, i.e. element-wise multiplication, the Manhattan difference [i.e. the absolute differences of corresponding features (Mueller and Thyagarajan, 2016)] and concatenation. The Manhattan difference consistently obtains better performance, considering the small values of the input features and the asymmetry of the captured protein relations. Results. The prediction accuracy and fold changes over the zero rule baseline are reported in Table 4. Note that since the multi-class prediction task is much more challenging than the binary prediction task, it is expected to observe lower accuracy and longer training time (Table 7) than that reported in the previous experiment. Among all the baselines using explicit features, the CTD-based models perform better than the AC-based ones. CTD descriptors seek to cover both continuous and discontinuous interaction information (Yang ), which potentially better discriminate among PPI types.
Table 4.

Accuracy (%) and fold changes over zero rule for PPI interaction type prediction on two STRING datasets based on 10-fold cross-validation

FeaturesN/A
AC
CTD
Embedded raw seqs
MethodsRandZero ruleSVMRFAdaBoostkNNLogisticSVMRFAdaBoostkNNLogisticSCNNSRGRU PIPR
SHS27k14.2816.7033.1744.8228.6735.4425.4735.5645.7631.8135.5630.5755.5451.06 59.56
(fold×)1.00×1.99×2.68×1.72×2.12×1.52×2.13×2.74×1.90×2.13×1.83×3.33×3.06× 3.57×
SHS148k14.2816.2128.1736.0127.8733.8124.9631.3736.6529.6733.1326.9655.2954.05 61.91
(fold×)1.00×1.74×2.22×1.72×2.09×1.54×1.94×2.26×1.83×2.04×1.66×3.41×3.33× 3.82×

Each boldfaced number indicates the best of the corresponding metric.

Accuracy (%) and fold changes over zero rule for PPI interaction type prediction on two STRING datasets based on 10-fold cross-validation Each boldfaced number indicates the best of the corresponding metric. The best baseline using Random Forest thereof achieves satisfactory results by more than doubling the accuracy of zero rule on the smaller SHS27k dataset. However, on the larger SHS148k dataset, the accuracy of these explicit-feature-based models is notably impaired. We hypothesize that such predefined explicit features are not representative enough to distinguish the PPI types. On the other hand, the deep-learning-based approaches do not need to explicitly utilize these features, and perform consistently well in both settings. The raw sequence information is sufficient for these approaches to drastically outperform the Random Forest by at least 5.30% in accuracy on SHS27k and 17.40% in accuracy on SHS148k. SCNN thereof outperforms SRGRU by 4.48 and 1.24% in accuracy on SHS27k and SHS148k, respectively. This implies that the local interacting features are relatively more deterministic than contextualized and sequential features on this task. The results by the residual RCNN-based framework are very promising, as it outperforms SCNN by 4.02% and 6.62% in accuracy on SHS27k and SHS148k, respectively. It also remarkably outperforms the best explicit-feature-based baselines on the two datasets by 13.80 and 25.26% in accuracy, and more than 3.5 of fold changes over the zero rule on both datasets.

4.4 Binding affinity estimation

Lastly, we evaluate PIPR for binding affinity estimation using the SKEMPI dataset. We employ the mean squared loss variant of PIPR to address this regression task. Since the lengths of protein sequences in SKEMPI are much shorter than those in the other datasets, we accordingly reduce the occurrences of RCNN units to three, while other configurations remain unchanged. For baselines, we compare against several regression models based on the AC and CTD features, which include Bayesian Ridge regressor (BR), SVM, AdaBoost with decision tree regressors and Random Forest regressor. The corresponding features for two sequences are again combined via the Manhattan difference. We also modify SCNN and SRGRU to their mean squared loss variants, in which we reduce the layers in the same way of RCNN. Evaluation protocol. We aggregate three metrics through 10-fold CV, i.e. mean squared error (MSE), mean absolute error (MAE) and Pearson’s correlation coefficient (Corr). These are three commonly reported metrics for regression tasks, for which lower MSE and MAE as well as higher Corr indicate better performance. In the CV process, we normalize the affinity values of the SKEMPI dataset to via min–max re-scaling. (This is due to that we use sigmoid function to smooth the output of the regressor. Note that this process does not affect correlation, while MSE, MAE and the original affinity scores can be easily re-scaled back.). Results. Table 5 reports the results for this experiment. It is noteworthy that, one single change of amino acid can lead to a drastic effect on binding affinity. While such subtle changes are difficult to be reflected by the explicit features, the deep-learning-based methods can competently capture such changes from the raw sequences. Our RCNN-based framework again offers the best performance among the deep-learning-based approaches, and significantly outperforms the best baseline (CTD-based Random Forest) by offering a 0.233 increase in Corr, as well as remarkably lower MSE and MAE. Figure 3 demonstrates an example of the effect of changing an amino acid in a protein complex. Tyrosine at position 61 of Chymotrypsin inhibitor 2 (Chain I) is substituted with Alanine, causing the neighboring region of Subtilisin BPN’ precursor (Chain E) to relax. The binding affinity (k) changes from 2.24E-12 to 2.70E-10, which is validly captured by PIPR. While our experiment is conducted on a relatively small dataset, we seek to extend our PIPR framework to a more generalized solution for binding affinity estimation, once a larger and more heterogeneous corpus is available.
Table 5.

Results for binding affinity prediction on the SKEMPI dataset

FeaturesAC
CTD
Embedded raw seqs
MethodsBRSVMRFAdaBoostBRSVMRFAdaBoostSCNNSRGRU PIPR
MSE (×102)1.702.201.771.981.861.841.491.840.870.95 0.63
MAE (×10-2)9.5611.819.8111.1510.2011.049.0610.696.497.08 5.48
Corr 0.5640.3530.5460.4510.5010.5010.6400.5080.8310.812 0.873

Note: Each measurement is an average of the test sets over 10-fold cross-validation.

Each boldfaced number indicates the best of the corresponding metric.

Fig. 3.

Mutation effects on structure and binding affinity. The blue entity is Subtilisin BPN’ precursor (Chain E), and the red entity is Chymotrypsin inhibitor (Chain I). The mutation is highlighted in yellow. The wild-type (1TM1) and mutant (1TO1) complexes are retrieved from PDB

Results for binding affinity prediction on the SKEMPI dataset Note: Each measurement is an average of the test sets over 10-fold cross-validation. Each boldfaced number indicates the best of the corresponding metric. Mutation effects on structure and binding affinity. The blue entity is Subtilisin BPN’ precursor (Chain E), and the red entity is Chymotrypsin inhibitor (Chain I). The mutation is highlighted in yellow. The wild-type (1TM1) and mutant (1TO1) complexes are retrieved from PDB

4.5 Amino acid embeddings

We further investigate the settings of amino acid embeddings in this subsection. Each amino acid is represented by a vector of numerical values that describe its relative physicochemical properties. The first part of the embedding vector, which measures the co-occurrence similarity of the amino acids in protein sequences, is empirically set as a five-dimensional vector. is obtained by pre-training the Skip-Gram model on all 8000 sequences from our largest STRING dataset, SHS148k, using a context window size of seven and a negative sampling size of five. The second part contains a seven-dimensional vector, , which describes the categorization of electrostaticity and hydrophobicity for the amino acid. We examine the performance of using each part individually, as well as the performance of combining them as used in our framework. In addition, we include a naïve one-hot vector representation, which does not consider the relatedness of amino acids and treats each of them independently. Table 6 shows that, once we remove either of the two parts of the proposed embedding, the performance of the model slightly drops. Meanwhile, the proposed pre-trained embeddings lead to noticeably better performance of the model than adopting the naïve one-hot encodings of the canonical amino acids. This pre-training process completes in 8 s on a commodity workstation as shown in Table 7. This is a one-time effort that can be reused on different tasks and datasets.
Table 6.

Comparison of amino acid representations based on binary prediction

[ac,aph] ac only aph onlyOne-hot
Dimension125720
Accuracy 97.09 96.6796.0396.11
Precision 97.00 96.3595.9196.34
F1-score 97.09 96.5196.0896.10
Comparison of amino acid representations based on binary prediction Run-time of training embeddings and different prediction tasks

4.6 Run-time analysis

All of the experiments are conducted on one NVIDIA GeForce GTX 1080 Ti GPU. We report the training time for each experiment, as well as for the amino acid embedding in Table 7. For each experiment, we calculate the average training time over either 5-fold (Yeast dataset) or 10-fold (others) CV. In both binary and multi-class predictions, the training time increases along with the increased number of training cases. The regression estimation generally requires more iterations per training case to converge than classification tasks. Thus, with much fewer cases, the training time on SKEMPI for affinity estimation is more than that on the Yeast dataset for binary prediction.

5 Conclusion

In this paper, we propose a novel end-to-end framework for PPI prediction based on the amino acid sequences. Our proposed framework, PIPR, employs a residual RCNN, which provides an automatic multi-granular feature selection mechanism to capture both local significant features and sequential features from the primary protein sequences. By incorporating the RCNN in a Siamese-based learning architecture, the framework captures effectively the mutual influence of protein pairs, and generalizes well to address different PPI prediction tasks without the need for predefined features. Extensive experimental evaluations on five datasets show promising performance of our framework on three challenging PPI prediction tasks. This also leads to significant amelioration over various baselines. Experiments on datasets of different sizes also demonstrate satisfactory scalability of the framework. For future work, one important direction is to apply the PIPR framework to other sequence-based inference tasks in bioinformatics, such as modeling RNA and protein interactions. We also seek to incorporate attention mechanisms (Vaswani ) to help pinpoint interaction sites on protein sequences, and apply PIPR to predict confidence of interactions in the form of ordinal regression. Since PIPR has alleviated any costly domain-invariant feature engineering process, how to extend PIPR with transfer learning based domain adaptation for different species is another meaningful direction. Click here for additional data file.
  38 in total

1.  The Protein Data Bank.

Authors:  H M Berman; J Westbrook; Z Feng; G Gilliland; T N Bhat; H Weissig; I N Shindyalov; P E Bourne
Journal:  Nucleic Acids Res       Date:  2000-01-01       Impact factor: 16.971

2.  A Bayesian networks approach for predicting protein-protein interactions from genomic data.

Authors:  Ronald Jansen; Haiyuan Yu; Dov Greenbaum; Yuval Kluger; Nevan J Krogan; Sambath Chung; Andrew Emili; Michael Snyder; Jack F Greenblatt; Mark Gerstein
Journal:  Science       Date:  2003-10-17       Impact factor: 47.728

3.  The Database of Interacting Proteins: 2004 update.

Authors:  Lukasz Salwinski; Christopher S Miller; Adam J Smith; Frank K Pettit; James U Bowie; David Eisenberg
Journal:  Nucleic Acids Res       Date:  2004-01-01       Impact factor: 16.971

4.  Cd-hit: a fast program for clustering and comparing large sets of protein or nucleotide sequences.

Authors:  Weizhong Li; Adam Godzik
Journal:  Bioinformatics       Date:  2006-05-26       Impact factor: 6.937

Review 5.  Computational prediction of protein-protein interactions.

Authors:  Lucy Skrabanek; Harpreet K Saini; Gary D Bader; Anton J Enright
Journal:  Mol Biotechnol       Date:  2007-08-14       Impact factor: 2.695

6.  Predicting protein-protein interactions based only on sequences information.

Authors:  Juwen Shen; Jian Zhang; Xiaomin Luo; Weiliang Zhu; Kunqian Yu; Kaixian Chen; Yixue Li; Hualiang Jiang
Journal:  Proc Natl Acad Sci U S A       Date:  2007-03-05       Impact factor: 11.205

7.  Functional organization of the yeast proteome by systematic analysis of protein complexes.

Authors:  Anne-Claude Gavin; Markus Bösche; Roland Krause; Paola Grandi; Martina Marzioch; Andreas Bauer; Jörg Schultz; Jens M Rick; Anne-Marie Michon; Cristina-Maria Cruciat; Marita Remor; Christian Höfert; Malgorzata Schelder; Miro Brajenovic; Heinz Ruffner; Alejandro Merino; Karin Klein; Manuela Hudak; David Dickson; Tatjana Rudi; Volker Gnau; Angela Bauch; Sonja Bastuck; Bettina Huhse; Christina Leutwein; Marie-Anne Heurtier; Richard R Copley; Angela Edelmann; Erich Querfurth; Vladimir Rybin; Gerard Drewes; Manfred Raida; Tewis Bouwmeester; Peer Bork; Bertrand Seraphin; Bernhard Kuster; Gitte Neubauer; Giulio Superti-Furga
Journal:  Nature       Date:  2002-01-10       Impact factor: 49.962

8.  Systematic identification of protein complexes in Saccharomyces cerevisiae by mass spectrometry.

Authors:  Yuen Ho; Albrecht Gruhler; Adrian Heilbut; Gary D Bader; Lynda Moore; Sally-Lin Adams; Anna Millar; Paul Taylor; Keiryn Bennett; Kelly Boutilier; Lingyun Yang; Cheryl Wolting; Ian Donaldson; Søren Schandorff; Juanita Shewnarane; Mai Vo; Joanne Taggart; Marilyn Goudreault; Brenda Muskat; Cris Alfarano; Danielle Dewar; Zhen Lin; Katerina Michalickova; Andrew R Willems; Holly Sassi; Peter A Nielsen; Karina J Rasmussen; Jens R Andersen; Lene E Johansen; Lykke H Hansen; Hans Jespersen; Alexandre Podtelejnikov; Eva Nielsen; Janne Crawford; Vibeke Poulsen; Birgitte D Sørensen; Jesper Matthiesen; Ronald C Hendrickson; Frank Gleeson; Tony Pawson; Michael F Moran; Daniel Durocher; Matthias Mann; Christopher W V Hogue; Daniel Figeys; Mike Tyers
Journal:  Nature       Date:  2002-01-10       Impact factor: 49.962

9.  NOXclass: prediction of protein-protein interaction types.

Authors:  Hongbo Zhu; Francisco S Domingues; Ingolf Sommer; Thomas Lengauer
Journal:  BMC Bioinformatics       Date:  2006-01-19       Impact factor: 3.169

10.  Using support vector machine combined with auto covariance to predict protein-protein interactions from protein sequences.

Authors:  Yanzhi Guo; Lezheng Yu; Zhining Wen; Menglong Li
Journal:  Nucleic Acids Res       Date:  2008-04-04       Impact factor: 16.971

View more
  26 in total

1.  Current challenges for unseen-epitope TCR interaction prediction and a new perspective derived from image classification.

Authors:  Pieter Moris; Joey De Pauw; Anna Postovskaya; Sofie Gielis; Nicolas De Neuter; Wout Bittremieux; Benson Ogunjimi; Kris Laukens; Pieter Meysman
Journal:  Brief Bioinform       Date:  2021-07-20       Impact factor: 11.622

Review 2.  Computational Tools and Strategies to Develop Peptide-Based Inhibitors of Protein-Protein Interactions.

Authors:  Maxence Delaunay; Tâp Ha-Duong
Journal:  Methods Mol Biol       Date:  2022

3.  D-SCRIPT translates genome to phenome with sequence-based, structure-aware, genome-scale predictions of protein-protein interactions.

Authors:  Samuel Sledzieski; Rohit Singh; Lenore Cowen; Bonnie Berger
Journal:  Cell Syst       Date:  2021-10-09       Impact factor: 11.091

4.  Computational Methods and Deep Learning for Elucidating Protein Interaction Networks.

Authors:  Dhvani Sandip Vora; Yogesh Kalakoti; Durai Sundar
Journal:  Methods Mol Biol       Date:  2023

5.  Topsy-Turvy: integrating a global view into sequence-based PPI prediction.

Authors:  Rohit Singh; Kapil Devkota; Samuel Sledzieski; Bonnie Berger; Lenore Cowen
Journal:  Bioinformatics       Date:  2022-06-24       Impact factor: 6.931

6.  Multi-view heterogeneous molecular network representation learning for protein-protein interaction prediction.

Authors:  Xiao-Rui Su; Lun Hu; Zhu-Hong You; Peng-Wei Hu; Bo-Wei Zhao
Journal:  BMC Bioinformatics       Date:  2022-06-16       Impact factor: 3.307

7.  Assessing sequence-based protein-protein interaction predictors for use in therapeutic peptide engineering.

Authors:  François Charih; Kyle K Biggar; James R Green
Journal:  Sci Rep       Date:  2022-06-10       Impact factor: 4.996

8.  Protein-protein interaction and non-interaction predictions using gene sequence natural vector.

Authors:  Nan Zhao; Maji Zhuo; Kun Tian; Xinqi Gong
Journal:  Commun Biol       Date:  2022-07-02

9.  Prediction of Protein-Protein Interactions with Local Weight-Sharing Mechanism in Deep Learning.

Authors:  Lei Yang; Yukun Han; Huixue Zhang; Wenlong Li; Yu Dai
Journal:  Biomed Res Int       Date:  2020-06-13       Impact factor: 3.411

10.  AutoPPI: An Ensemble of Deep Autoencoders for Protein-Protein Interaction Prediction.

Authors:  Gabriela Czibula; Alexandra-Ioana Albu; Maria Iuliana Bocicor; Camelia Chira
Journal:  Entropy (Basel)       Date:  2021-05-21       Impact factor: 2.524

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.