Literature DB >> 30713552

A Novel Protein Subcellular Localization Method With CNN-XGBoost Model for Alzheimer's Disease.

Long Pang1, Junjie Wang2, Lingling Zhao3, Chunyu Wang2, Hui Zhan2.   

Abstract

The disorder distribution of protein in the compartment or organelle leads to many human diseases, including neurodegenerative diseases such as Alzheimer's disease. The prediction of protein subcellular localization play important roles in the understanding of the mechanism of protein function, pathogenes and disease therapy. This paper proposes a novel subcellular localization method by integrating the Convolutional Neural Network (CNN) and eXtreme Gradient Boosting (XGBoost), where CNN acts as a feature extractor to automatically obtain features from the original sequence information and a XGBoost classifier as a recognizer to identify the protein subcellular localization based on the output of the CNN. Experiments are implemented on three protein datasets. The results prove that the CNN-XGBoost method performs better than the general protein subcellular localization methods.

Entities:  

Keywords:  Conventional Neural Network (CNN); XGBoost; deep learning (DL); machine learning; protein subcellular localization

Year:  2019        PMID: 30713552      PMCID: PMC6345701          DOI: 10.3389/fgene.2018.00751

Source DB:  PubMed          Journal:  Front Genet        ISSN: 1664-8021            Impact factor:   4.599


1. Introduction

The study of neurodegenerative diseases, specifically the Alzheimer's disease(AD) has gained great attention and been addressed widely (Cai et al., 2013; Hu et al., 2017a,b,c). The abnormalities and disorder distribution the compartment or organelle of tau protein and the beta-amyloid protein have been considered to contribute to the pathogenesis of AD. Protein subcellular localization prediction is an essential task in bioinformatics and plays import roles in the further understanding of the relationship among protein locations, their function exhibition, and nosogenesis (Liu et al., 2015; Cheng et al., 2016a, 2018a). Related predictive tools typically use the amino acid sequence information of the protein itself as input to output predicted protein cell sublocalization. It provides information on protein function and gene annotation to aid in the identification of drug targets. The two commonly used methods are: (1) homology-based method and (2) machine learning based method (Wu and Krishnan, 2011; Wu et al., 2014; Zeng et al., 2014; Cheng et al., 2017). The homology-based method highly depends on the homology of protein sequences, and therefore performs worse for low protein sequence similarity (Wei et al., 2016; Cheng et al., 2018b). The machine learning based methods usually extract some features from the amino acid sequence of the protein (Cheng et al., 2016b; Hu et al., 2018), convert the sequence into a numerical vector, and then use a machine learning model to predict. For example, the most widely used WoLF PSORT software for eukaryotic proteins, characterized by the amino acid composition of the protein, gives the cellular sublocalization of the 32 proteins most similar to the input protein using the k-nearest neighbor algorithm (Horton et al., 2007). There also exist similar methods like BaCelLo (Pierleoni et al., 2006), YLoc (Briesemeister et al., 2010), iLoc-Hum (Chou et al., 2012), and Hum-mPLoc 3.0 (Zhou et al., 2016). We believe that existing forecasting methods also have some room for improvement. First, the extracted sequence characteristics may not fully reflect the properties of the protein associated with the training task. Second, the current predictions only use information about the protein itself, without considering the interaction between proteins. In recent years, deep learning has been proven to be a very powerful method by researchers in many fields (LeCun et al., 2015; Xu et al., 2017), like computer vision and natural language processing (Krizhevsky et al., 2012; Mikolov et al., 2013; Sutskever et al., 2014). CNN is an efficient deep learning method due to it can learn high-level features with neural networks. Recently, it also has attracted attentions from researchers and practitioners in bioinformatics. A prediction tool “DeepLoc” (Almagro Armenteros et al., 2017) based on deep learning was proposed with the end-to-end sequence-based model integrated recurrent neural networks (RNNs) with long short-term memory(LSTM) cells, attention models and convolutional neural networks(CNNs), and achieved a better accuracy compared with the traditional machine learning methods. However, the model structure is of high complication, sequentially has too many hyper-parameters to train. Moreover, the proteins in the dataset they constructed have been found to be highly homologous and therefore might provide an overly optimistic model evaluation (Gudenas, 2018). In addition, DeepLoc considers only one possible label for each protein, whereas the protein subcellular location belongs to a multi-label multi-class problem in general. In this work, we propose a new framework for protein subcellular localization prediction by combining CNN and XGBoost. As an outstanding classifier and feature extractor, CNNs have achieved great success, especially in the field of image recognization. For the protein sequence, CNNs have ability to detect short motifs in the input sequence irrespectively of where they occur and automatically extract features from the original protein sequences. Inspired by this advantage, we also exploit CNN as the feature extractor but a new classifier XGBoost to replace the traditional classifiers connected like the soft-max classifier, since they can not well understand the extracted feature by CNN. XGBoost is an efficient implementation of gradient boosted decision trees (GBDT) due to its block structure to support the parallelization of tree construction. In GBDT, gradient boosting refers to a kind of ensemble technique creating new models to predict the residuals or errors of prior models and making the final decision by the summing up the predictions from all models. Meantime, gradient descent algorithm is also exploited to minimize the loss when adding new models. The main contribution of our work includes the following aspects: We propose a new CNN-XGBoost model for prediction of the protein subcellular localization. The high-level features of protein sequence can be learned by a CNN that can be used by XGBoost classifier for prediction the localization of the subcellular of proteins. The experiments conducted on four real datasets consisting of protein sequences show that the proposed method achieves competitive performance.

2. Methods

In this paper, we propose a novel protein subcellular localization method by integrating the CNN and the XGBoost as a new model for possible application in the pathogenes verification of Alzheimer's disease. The general concept of CNN-XGBoost model is to add an XGboost after the feature layer of a CNN and replace the output layer of the CNN. Our CNN-XGBoost model can automatically extract featutue from the protein sequences and provides more precise localization results. Figure 1 illustrates the whole structure of the CNN-XGBoost model for protein subcellular localization.
Figure 1

The framework of the CNN-XGBoost based protein subcellular location predictor.

The framework of the CNN-XGBoost based protein subcellular location predictor.

2.1. Convolutional Neural Network

In the field of image analysis, the mask (or filter, or kernel) is an important construct. A convolution is an operation involving an initial image and the mask. The operation is equivalent to flipping the mask both vertically and horizontally and then visually placing it over each pixel in turn. The output is the sum over a pixel-wise product of the mask and the sub-image. Masks are usually symmetric, so flipping is unnecessary. Recall from signal processing, the convolution between two f and g is given by the following equation. In image processing, a convolution between an image I and kernel K of size d × d centered at a given pixel (x, y) is defined as, Convolutional neural networks are a family of neural network architectures having at least one convolutional layer. LeNet is the original CNN network architecture bearing the name of Yann Lecun. Its architecture can be written as, In this architecture, convolutional layer is the cornerstone of the CNN, which is a hidden layer where a square grid of weights is convolved with the input, just like an image mask. The output of the convolutional layer is akin to a convolved image. Next, the non-linear activation function, ReLu (REctified Linear Unit), is applied to zero-out any negative values. To reduce the dimension of the feature extracted from the convolutional layer, there is a pooling layer emulating downsampling. In general, each group of four values or pixels is replaced by the maximum (sometimes the mean) of the four, leaving a single most intense pixel. This pooling method is known as max pooling. This sequence of CONV->RELU->POOLlayers may be repeated multiple times to create a deep architecture. Finally, a few fully-connected layers round off the architecture. Though it seems far more sophisticated than a MLP, it can be shown that a CNN can be represented as a classical fully-connected neural network. For example, a convolutional layer can be represented as a sparse fully-connected layer. Various techniques have been developed for training these vast models, for example momentum optimizers, weight initialization, batch normalization, and dropout. Convolutional Neural Networks are the current state-of-the-art in many computer vision tasks. In addition to image classification, their great success has attracted wide attention in many fields. It has been found that using a pre-trained CNN as a general-purpose feature extractor for a simple linear model can yield significant improvements over even the most meticulously hand-crafted feature engineering. The protein subcellular localization problem can be viewed as a multi-label multi-class classification task. Unlike the ordinary deep learning methods for multi-classification problems, in our method, we need to change the loss function. The most intuitive way is to extend the cross-entropy loss. The cross-entropy loss function is defined by where Θ denotes the parameters of CNN model, is a set that contains the relevant localization of protein i and is the result for protein i on localization j, through a softmax activation: Instead of using the cross-entropy loss function, the binary cross-entropy loss (BCE) over sigmoid activation has shown better performance when applied into multi-label task. The binary cross-entropy loss is where

2.2. Tree Boosting and XGBoost

Tree boosting is a learning method to enhance the classification ability of weak classifiers by iteratively adding new decision trees to the ensembles of decision trees. Let denotes a dataset with n classes and m feature. Then the prediction of a tree boosting for a (x, y) is given by where g(x) = w(x) is the prediction of the j-th decision tree with leaf weights w on a datapoint x, and M is the number of members in the ensemble. It is well-known that the decision tree tends to overfit when the decision tree is fully grown. Thus, the set prediction function of decision trees g can be learned by minimizing the objective function where l(y, ŷ) is a term which measures the goodness of the prediction ŷ and the object y. Ω(g) is a regularization term that does not depend on the data. XGBoost implements a parallel tree boosting in a fast and accurate way. In XGBoost, the regularization function is chosen to be with γ and λ regularization parameters that must be chosen appropriately. Notice this regularization penalizes both large weights on the leaves (similar to L2-regularization) and has large partitions. As mentioned above, the tree boosting iteratively enlarges the ensemble of decision trees, then the prediction of the t-th iteration can be defined as The objective function (7) at step t can be modified as Apply a Taylor expansion on the objective function (10) to second order and then the final objective function at step t can be approximated as where Let j:I = {i:q(x) = j} denotes the set of point x mapped to leaf, and . Then we can rewrite the as To find the optimal weight w of leaf j for a fixed tree structure, q(x) can be obtained by applying the following equation plugging back into gives It is clear that measures the in-sample performance of g and we should find the decision tree that minimizes this value. However, in practice, this is impossible to enumerate all possible trees over the data and find the tree which can minimize . Instead, an approximate greedy algorithm runs to optimize one level of the tree at a time by trying to find optimal splits of the data, leading to a tree with a local minimum of , which is then added to the ensemble. For the multi-label multi-class classification problem, we utilize XGBoost as classifiers and adopt the binary relevance strategy (Boutell et al., 2004) to construct m binary classifiers.

2.3. CNN-XGBoost Model

Figure 1 gives the overall structure of the CNN-XGBoost model for protein subcellular location prediction. The input of the model is a one-dimensional vector and constructed by the position specific scoring matrices (PSSM) and proteins interaction scoring matrix which are extracted from STRING and GO terms semantic similarities. On this basis, a protein can be expressed as L × 1 vector (L is the number of sequences in training set), analog image data equivalent to a protein is a one-dimensional “image” with 1 channels. So the input is a L × 1 matrix. After obtaining the proper feature representations by the trained CNN, compared with the classic CNN, our CNN-XGBoost model replaces the soft-max layer of CNN with XGBoost to predict the localization of subcellular of proteins, which enables features automatically obtained from input and provides more precise and efficient classification.

3. Results

3.1. Dataset

To verify the performance of our method, we employ three protein datasets: the Hum-mPloc3.0, the BaCelLo animals, and the Hoglund. Table 1 gives the details of these datasets. The train set of Hum-mPloc 3.0 consists of 3,122 proteins and 1,023 proteins own more than one label. The test set of Hum-mPloc 3.0 consists of 379 proteins, among which 120 proteins belong to multi-label proteins. Each protein in Hum-mPloc 3.0 is assigned at least one label of 12 subcellular locations (Centrosome, Cytoplasm, Cytoskeleton, Endoplasmic reticulum, Endosome, Extracellular, Golgi apparatus, Lysosome, Mitochondrion, Nucleus, Peroxisome, and Plasma membrane).
Table 1

Dataset Summary.

Hum-mLoc 3.0BaCelLoHoglund
TrainingTestingTrainingTestingTrainingTesting
No. Proteins3,1263792,5975765,959158
No. Labels4,2295412,5975765,959158
No.Locations1246
Dataset Summary. For the BaCelLo dataset, there are four subcellular locations: Cytoplasm, Mitochondrion, Nucleus, and Secreted. The size of the training set is set to 2,597 and the testing set consists of 576 proteins. All the proteins of BaCelLo dataset are of a single label. In the Hoglund dataset, the training set includes nine subcellular locations (Nucleus, Cytoplasm, Mitochondrion, Endoplasmic reticulum, Golgi apparatus, Peroxisome, Plasma membrane, Extracellular space, Lysosome, and Vacuole), and the test consists of 158 proteins with six subcellular locations (Endoplasmic reticulum, Golgi apparatus, Peroxisome, Plasma membrane, Extracellular space, and Lysosome).

3.2. Measurements

A widely-applied method for evaluating a mutli-label multi-class classifier is to compute the ACC and F1 values. ACC is the average of ACC of all proteins in the testing set, calculated for protein x is where TP, FP, and FN are true positive, false positive, and false negative, respectively. The F1 score considers both the harmonic mean of precision and recall of subcellular location y, defined as follows: where T and P are the set of proteins for true location y and the set of proteins for predicted locations y respectively.

3.3. Results and Discussions

To verify the performance of our approach, some typical protein subcellular location tools including Hum-mPLoc 3.0 (Zhou et al., 2016), YLoc+ (Briesemeister et al., 2010), iLoc-Hum (Chou et al., 2012), WegoLoc (Chi and Nam, 2012), mLASSO-Hum (Wan et al., 2015), and PSL-Recommender (Jamali et al., 2018) were compared to our method. The F1 score and ACC for each subcellular localization are summarized in Table 2 and Figure 2 for Hum-mploc 3.0 dataset. As seen in Table 2 and Figure 2, the CNN-XGBoost outperforms the mean value of F1 score and ACC of all other methods. Also, in 7 out of 12 subcellular locations, CNN-XGBoost has the best performance among all the methods while in the other three locations it has the second best performance. It is only in centrosome and endosome that CNN-XGBoost shows unsatisfactory results. As seen in Table 3, the CNN-XGBoost slightly outperforms the second best method by both mean F1 score and ACC.
Table 2

Comparision of CNN-XGBoost on Hum-mPloc 3.0 dataset with other methods.

LocationiLoc-HumanWegoLocmLASSO-HumHum-mLoc 3.0PSL-RecommenderCNN-XGBoost
prereF1prereF1prereF1prereF1prereF1prereF1
Centrosome0000.750.140.230.590.590.590.750.550.630.940.750.830.790.500.61
Cytoplasm0.50.540.520.690.530.600.930.510.660.760.730.740.790.810.800.850.890.87
Cytoskeleton0000.320.340.330.90.220.350.80.680.740.930.770.840.890.800.85
ER0000.730.20.310.740.490.590.830.370.510.90.70.790.970.710.82
Endosome0000.250.070.110.380.20.260.580.470.520.570.370.450.800.270.40
Extracellular0.620.620.620.670.770.710.160.690.260.50.460.480.660.710.680.800.620.70
Golgi apparatus0.60.30.40.60.150.240.720.650.680.690.450.550.880.610.720.800.600.69
Lysosome0.50.130.20.20.130.150.550.750.630.710.630.6710.550.711.000.750.86
Mitochondrion0.950.330.490.790.730.760.830.880.850.780.750.760.920.880.900.960.800.87
Nucleus0.540.70.610.650.640.640.850.70.760.750.710.730.810.920.870.830.910.87
Peroxisome10.50.670.510.670.2910.44111111111
Plasma membrane0.420.330.370.440.530.480.580.560.570.650.440.520.780.740.760.890.750.81
ACC-mean0.410.500.650.630.770.78
F1-mean0.320.440.560.650.780.80

The bold marks the first best result and the underline marks the second best result.

Figure 2

The accuracy comparison on the Hum-mPloc 3.0 data set.

Table 3

Comparison of CNN-XGBoost ACC/F1-mean on other proteins datasets with other methods.

BaCelLoHoglund
MultiLoc2-LowRes0.73/0.76
MultiLoc2-HighRes0.68/0.710.57/0.41
BaCelLo0.64/0.66
Hum-mPloc 3.00.86/0.840.64/0.59
PSL-Recommender0.94/0.920.92/0.90
CNN-XGBoost0.94/0.940.94/0.92

The bold marks the first best result and the underline marks the second best result.

Comparision of CNN-XGBoost on Hum-mPloc 3.0 dataset with other methods. The bold marks the first best result and the underline marks the second best result. The accuracy comparison on the Hum-mPloc 3.0 data set. Comparison of CNN-XGBoost ACC/F1-mean on other proteins datasets with other methods. The bold marks the first best result and the underline marks the second best result. In addition, we also evaluated our method on the DeepLoc dataset, compared to the DeepLoc, our method provides slightly better prediction with significantly lighter model, meanwhile, it is known that DeepLoc can not handle multilabel multiclass problem, whereas our method still shows outstanding performance.

4. Conclusions

In order to make balance of the classification performance and the complexity when training the model for the protein subcellular location in Alzheimer's disease, this paper proposes a prediction framework integrating CNN and XGBoost, taking advantage of the outstanding ability of feature expression of CNN, and the good classification performance of XGBoost. Experiments are implemented on the Hum-mPloc3.0, the BaCelLo animals, and the Hoglund database, and the results demonstrate that the new method outperforms the typical machine learning based tools. Further work will focus on the verification of our model on more datasets, especially the datasets related to Alzheimer's disease, and the optimization of the structure of CNN utilized in the model.

Author Contributions

All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
  22 in total

1.  iLoc-Hum: using the accumulation-label scale to predict subcellular locations of human proteins with both single and multiple sites.

Authors:  Kuo-Chen Chou; Zhi-Cheng Wu; Xuan Xiao
Journal:  Mol Biosyst       Date:  2011-12-01

2.  WegoLoc: accurate prediction of protein subcellular localization using weighted Gene Ontology terms.

Authors:  Sang-Mun Chi; Dougu Nam
Journal:  Bioinformatics       Date:  2012-01-31       Impact factor: 6.937

3.  Implementation of Arithmetic Operations With Time-Free Spiking Neural P Systems.

Authors:  Xiangrong Liu; Ziming Li; Juan Liu; Logan Liu; Xiangxiang Zeng
Journal:  IEEE Trans Nanobioscience       Date:  2015-09       Impact factor: 2.935

4.  BaCelLo: a balanced subcellular localization predictor.

Authors:  Andrea Pierleoni; Pier Luigi Martelli; Piero Fariselli; Rita Casadio
Journal:  Bioinformatics       Date:  2006-07-15       Impact factor: 6.937

Review 5.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

6.  mLASSO-Hum: A LASSO-based interpretable human-protein subcellular localization predictor.

Authors:  Shibiao Wan; Man-Wai Mak; Sun-Yuan Kung
Journal:  J Theor Biol       Date:  2015-07-09       Impact factor: 2.691

7.  Spiking neural P systems with thresholds.

Authors:  Xiangxiang Zeng; Xingyi Zhang; Tao Song; Linqiang Pan
Journal:  Neural Comput       Date:  2014-04-07       Impact factor: 2.026

8.  YLoc--an interpretable web server for predicting subcellular localization.

Authors:  Sebastian Briesemeister; Jörg Rahnenführer; Oliver Kohlbacher
Journal:  Nucleic Acids Res       Date:  2010-05-27       Impact factor: 16.971

9.  Knee joint vibration signal analysis with matching pursuit decomposition and dynamic weighted classifier fusion.

Authors:  Suxian Cai; Shanshan Yang; Fang Zheng; Meng Lu; Yunfeng Wu; Sridhar Krishnan
Journal:  Comput Math Methods Med       Date:  2013-03-12       Impact factor: 2.238

10.  WoLF PSORT: protein localization predictor.

Authors:  Paul Horton; Keun-Joon Park; Takeshi Obayashi; Naoya Fujita; Hajime Harada; C J Adams-Collier; Kenta Nakai
Journal:  Nucleic Acids Res       Date:  2007-05-21       Impact factor: 16.971

View more
  10 in total

1.  nhKcr: a new bioinformatics tool for predicting crotonylation sites on human nonhistone proteins based on deep learning.

Authors:  Yong-Zi Chen; Zhuo-Zhi Wang; Yanan Wang; Guoguang Ying; Zhen Chen; Jiangning Song
Journal:  Brief Bioinform       Date:  2021-11-05       Impact factor: 11.622

2.  Explainable Machine Learning Approach as a Tool to Understand Factors Used to Select the Refractive Surgery Technique on the Expert Level.

Authors:  Tae Keun Yoo; Ik Hee Ryu; Hannuy Choi; Jin Kuk Kim; In Sik Lee; Jung Sub Kim; Geunyoung Lee; Tyler Hyungtaek Rim
Journal:  Transl Vis Sci Technol       Date:  2020-02-12       Impact factor: 3.283

3.  Computational Identification and Analysis of Ubiquinone-Binding Proteins.

Authors:  Chang Lu; Wenjie Jiang; Hang Wang; Jinxiu Jiang; Zhiqiang Ma; Han Wang
Journal:  Cells       Date:  2020-02-24       Impact factor: 6.600

4.  A Comparative Analysis of Novel Deep Learning and Ensemble Learning Models to Predict the Allergenicity of Food Proteins.

Authors:  Liyang Wang; Dantong Niu; Xinjie Zhao; Xiaoya Wang; Mengzhen Hao; Huilian Che
Journal:  Foods       Date:  2021-04-09

5.  Use of Chou's 5-steps rule to predict the subcellular localization of gram-negative and gram-positive bacterial proteins by multi-label learning based on gene ontology annotation and profile alignment.

Authors:  Hafida Bouziane; Abdallah Chouarfia
Journal:  J Integr Bioinform       Date:  2020-06-29

Review 6.  Computational methods for protein localization prediction.

Authors:  Yuexu Jiang; Duolin Wang; Weiwei Wang; Dong Xu
Journal:  Comput Struct Biotechnol J       Date:  2021-10-19       Impact factor: 7.271

7.  Identification of Enzymes-specific Protein Domain Based on DDE, and Convolutional Neural Network.

Authors:  Rahu Sikander; Yuping Wang; Ali Ghulam; Xianjuan Wu
Journal:  Front Genet       Date:  2021-11-30       Impact factor: 4.599

8.  A hybrid machine learning/deep learning COVID-19 severity predictive model from CT images and clinical data.

Authors:  Matteo Chieregato; Fabio Frangiamore; Mauro Morassi; Claudia Baresi; Stefania Nici; Chiara Bassetti; Claudio Bnà; Marco Galelli
Journal:  Sci Rep       Date:  2022-03-14       Impact factor: 4.996

9.  An extended machine learning technique for polycystic ovary syndrome detection using ovary ultrasound image.

Authors:  Sayma Alam Suha; Muhammad Nazrul Islam
Journal:  Sci Rep       Date:  2022-10-12       Impact factor: 4.996

10.  DeepPred-SubMito: A Novel Submitochondrial Localization Predictor Based on Multi-Channel Convolutional Neural Network and Dataset Balancing Treatment.

Authors:  Xiao Wang; Yinping Jin; Qiuwen Zhang
Journal:  Int J Mol Sci       Date:  2020-08-09       Impact factor: 5.923

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.